diff --git a/.gitattributes b/.gitattributes
index b3d487871e5124acc2b73a0246a935b6cc85c74b..217cb4d692349619a6835f70d937b350311a92f9 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -173,3 +173,14 @@ runs/l2r50-baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs
runs/l2r50-baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r50-baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/l2r50-baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/l2r90-baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-1080/config.json b/runs/l2r90-baseline/checkpoint-1080/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-1080/generation_config.json b/runs/l2r90-baseline/checkpoint-1080/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-1080/model.safetensors b/runs/l2r90-baseline/checkpoint-1080/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6db7d84c02f1287ee824c056b6fd8e0a224beb4a
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c913a6ecc6b022c054063e727c029a2bff92dcf3208f8d9ba5977ff69619f497
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-1080/optimizer.pt b/runs/l2r90-baseline/checkpoint-1080/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d7c5a6c8c931aed33982d6c49c6e3aee526befe5
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4ebe66f5632d12006244fc83fa147e0ced7b1ca7f46ba26359253c46df20f8b1
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..691b2ea3e1b4f187dd5e6c0cda81e8858f6ac4cf
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:25e1b415517c5fe22089b6e5af2651408c03277402e7a2e6910a756d01bc912d
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8832e70f2d40e03d52900feac4951dac7fa3ec89
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5eac241b5f2b35e77529d15f1fe989e6cbe8bfc9e18a3b3e5d842da1f82ab4e2
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..cc6afe867703008555530c3c86fd6cf35ab7d7b1
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e2e72c528a5f7c6676fb4c9abd29d59ee303309497199e2b8bea743dbd9d88db
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4eb9917247f226d9cb7ed167e78dae92754ce06b
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40d0c6d939e35fba984de75edb8ae6eeab9a914372a646423e26bfff4aff7f03
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1080/scheduler.pt b/runs/l2r90-baseline/checkpoint-1080/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6460bf7d5e8aadff08654c5c446813d18e721d62
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bd8f1c3c3cf1427d3265ecb044cf76510643402bb89ce35df8f605375b066d14
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-1080/tokenizer.json b/runs/l2r90-baseline/checkpoint-1080/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1080/trainer_state.json b/runs/l2r90-baseline/checkpoint-1080/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..f067db6195192ed99b6fbd090569af5aaa385d7f
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/trainer_state.json
@@ -0,0 +1,7706 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.0,
+ "eval_steps": 500,
+ "global_step": 1080,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.604420041998336e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1080/training_args.bin b/runs/l2r90-baseline/checkpoint-1080/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1080/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-1440/config.json b/runs/l2r90-baseline/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-1440/generation_config.json b/runs/l2r90-baseline/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-1440/model.safetensors b/runs/l2r90-baseline/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ad8bc1822a865e8b9ac78873d3bab8e02d6f7e34
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bf794842ba70c1de9feac51edce37557dbe9c698c7740fbd60882205103446e9
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-1440/optimizer.pt b/runs/l2r90-baseline/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0c42825e08a66be2721996eb54e1f752527eb96f
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2d1ebc2f7d7a4cd34e2ea6acaa33dad13d10e049640e28cebc20e90657eca9b2
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..aeac82010ce2bacd37bd6af051976d38c4fe2812
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2b1dfed66aea4e59c5fbb97f6f7b9eedaa2102bc813da51af9a94ba9c42b0e8c
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f1f618778c4f7500b21ac62d13c3ef2bebccd4b9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:31bbc97cda656d11c5036ea59105511b2f07d7b2834b974ae356d2de6f8d922d
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5fdc074e62a1d8319b7a92b775a1095e5ea19aad
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f54eefd68dc84d612a90080eb0724dd4cdddd67659a3ca9aa7aeb336ba20c1fe
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ffe532f21731696b54618fdc406075cce9e04300
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:37d7701ca99aff6a37aa49666cef9609252548ec1ef723213fedc3c95ad648af
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1440/scheduler.pt b/runs/l2r90-baseline/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..59b8db4a485766e1d20d11a905a37774da0cec09
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2fe49e98b5d706cd9174dc7cf5e5cd5039ca03537426b9020792a5363720022f
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-1440/tokenizer.json b/runs/l2r90-baseline/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1440/trainer_state.json b/runs/l2r90-baseline/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..da1bc173cd768af9d922e1d8814377a8e81a2a22
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.139226722664448e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1440/training_args.bin b/runs/l2r90-baseline/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja b/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-1800/config.json b/runs/l2r90-baseline/checkpoint-1800/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-1800/generation_config.json b/runs/l2r90-baseline/checkpoint-1800/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-1800/model.safetensors b/runs/l2r90-baseline/checkpoint-1800/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c187dd9d2d4e7a90cbe7b8e403678f7314cb6a0e
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:18884bb3a15b51ad77d907f70762bbd36667561ebe02aca38a16037270e9ec38
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-1800/optimizer.pt b/runs/l2r90-baseline/checkpoint-1800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..46c0412ddcfa76dcac5435544f340cfe4b3ef7fb
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3194f6e151dbbffe602de2085a569217940541b1fea695aca41a784822b50e34
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0e1e82cfece0f9f4b960303264240de21a41f2b2
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8303cc5f635d6851ead500cf1dcb6114b8355d49a2ca28015d085bed26a36524
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4b06db675e31ec9653c89af439b51ab40f3cdab3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:99e1b9386d1bf84a8e2951517d247f0644fbc1877429b25df213b0870f58484c
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..88526073cc9c988c98d9074803bd3fbc38252bdd
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cb42dab9f2cc8d80099e5d46382913710a69bfab9566367bb8641baf78018e00
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..a9ec1a11b85512a95c1b62566c99d6bf9aa9c3ee
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:34a8598ceeb75c2fc78a1e5b43d2d5910250c2ea2784dd658d38680cdcec97b0
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-1800/scheduler.pt b/runs/l2r90-baseline/checkpoint-1800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..6e0ba2afe167e5831737cd4b431b166033f17fab
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ee9cab552a4cabfd465ec6f079c65fbe7304d4a3df75e2e2b3e2eb614e447c70
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-1800/tokenizer.json b/runs/l2r90-baseline/checkpoint-1800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1800/trainer_state.json b/runs/l2r90-baseline/checkpoint-1800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..686bf8e16127bbbdb9711812f708be77cdd88118
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/trainer_state.json
@@ -0,0 +1,12826 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.0,
+ "eval_steps": 500,
+ "global_step": 1800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.67403340333056e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-1800/training_args.bin b/runs/l2r90-baseline/checkpoint-1800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-1800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-2160/config.json b/runs/l2r90-baseline/checkpoint-2160/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-2160/generation_config.json b/runs/l2r90-baseline/checkpoint-2160/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-2160/model.safetensors b/runs/l2r90-baseline/checkpoint-2160/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..1d35be6fff00ef5890195bd2a404c4a944a2384d
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:54a768ac68dc73a4c2e6245fde5293c2acab5fa980caaf4327bea8514ec149bf
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-2160/optimizer.pt b/runs/l2r90-baseline/checkpoint-2160/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b9e2dbfc7f0043d744b78747b20804f613b986d5
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5c0d3fae8c96a300056716e6c1013391ead11e35338cbf8189dce430f33361c4
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..88d2248182b7fded85e1613b19e4b23df99807a8
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:08099d951fe225043bf5ed73f707acf33614689f66757b2a05711985ae68a49b
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f963be7f5cfeb7d06680588884ac6cd1c88f6843
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7de227da758404393630309e80adc97b4c158340b31e1f8ecfafc00d86ec519e
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..45ba75ff4448dedf7ab83e4464cb19a4bce5fc27
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:afa836b4fa8a014b4e440884fd4c2cf366a9b2474a622d72ac44bf2bedb57156
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..181c4a277a2b37c89f2d06083f8b4a7ca6902792
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6d94af1ab380b7f215a5ee3d3b50ab122adbd3dd0abfe48c29e83850925e0fa9
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2160/scheduler.pt b/runs/l2r90-baseline/checkpoint-2160/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..35aefe7762df83336d8bfddab210c09c85e6b3aa
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a7d6a0a5c94f665a32f55d7b034b4fc829f9f44c468d851ff5b24b22d8388e63
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-2160/tokenizer.json b/runs/l2r90-baseline/checkpoint-2160/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2160/trainer_state.json b/runs/l2r90-baseline/checkpoint-2160/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..5ab7f7d0ede6076cca695eb7007edc123b7fce0e
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/trainer_state.json
@@ -0,0 +1,15386 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.0,
+ "eval_steps": 500,
+ "global_step": 2160,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.078423500061035,
+ "eval_runtime": 57.3189,
+ "eval_samples_per_second": 42.604,
+ "eval_steps_per_second": 1.343,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.02624760940670967,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.261288642883301,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.03552987053990364,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.2682106494903564,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.040380869060754776,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.2813782691955566,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.03861293941736221,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.258193016052246,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.04091731086373329,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.3052220344543457,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.043269723653793335,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.2820217609405518,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.04580320790410042,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.312005043029785,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.047336459159851074,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.3217713832855225,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.04587360844016075,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.2869949340820312,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.04002532362937927,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.2879223823547363,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.0368693508207798,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.2983577251434326,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.037317775189876556,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.3062524795532227,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.03655700013041496,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.31072735786438,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.033466216176748276,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.2954514026641846,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.030785970389842987,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.27952241897583,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.029669342562556267,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.2613229751586914,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.033505845814943314,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.2987258434295654,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.032395437359809875,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.296708583831787,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.031322430819272995,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.3160252571105957,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.029345186427235603,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.307143211364746,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.026715675368905067,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.286012649536133,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.028592750430107117,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.2528645992279053,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.028861764818429947,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.2795748710632324,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.02609528973698616,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.326317071914673,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.027718454599380493,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.268110513687134,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.026382867246866226,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.294499397277832,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.027083583176136017,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.3202130794525146,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.02516467310488224,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.3485021591186523,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.024148069322109222,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.3099541664123535,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.024552857503294945,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.2885942459106445,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.02551114559173584,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.3232040405273438,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.025423647835850716,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.3097832202911377,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.023211417719721794,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.3075966835021973,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.024623898789286613,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.2917377948760986,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.02677663415670395,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.356265068054199,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.02673642337322235,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.314903736114502,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.024466225877404213,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.30912446975708,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.025370748713612556,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.363640069961548,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.026536114513874054,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.304851531982422,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.026536893099546432,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.342902898788452,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.02645307406783104,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.3164432048797607,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.02523060515522957,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.34376859664917,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.02590431459248066,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.3479061126708984,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.0230655986815691,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.2890982627868652,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.02403893694281578,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.345456600189209,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.02238907851278782,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.3471767902374268,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.021377569064497948,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.319232225418091,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.02272353693842888,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.316892623901367,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.02413824200630188,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.3450968265533447,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.02334682270884514,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.3640642166137695,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.021966520696878433,
+ "learning_rate": 0.000303,
+ "loss": 3.358555316925049,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.023984894156455994,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.358398914337158,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.023520268499851227,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.3224587440490723,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.0231285709887743,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.3671562671661377,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.022818170487880707,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.361027717590332,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.02280694991350174,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.2923383712768555,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.021362345665693283,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.3346734046936035,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.02372218668460846,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.354748249053955,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.022949764505028725,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.3857789039611816,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.021954255178570747,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.343330144882202,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.022438595071434975,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.328618049621582,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.021969925612211227,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.355971574783325,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.022706981748342514,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.3529577255249023,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.024083172902464867,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.348170757293701,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.023259354755282402,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.3474369049072266,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.02218157798051834,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.369053840637207,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.02305331453680992,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.3622641563415527,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.022796178236603737,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.351510524749756,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.021299205720424652,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.339406728744507,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.022786924615502357,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.3182196617126465,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.021729128435254097,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.383904218673706,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.024604283273220062,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 3.2912864685058594,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.0894880294799805,
+ "eval_runtime": 58.3967,
+ "eval_samples_per_second": 41.817,
+ "eval_steps_per_second": 1.319,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.02673362009227276,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.270986795425415,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.04036599025130272,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.2358999252319336,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.047250088304281235,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.2706189155578613,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.04642819985747337,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.236340284347534,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.039405327290296555,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.203136682510376,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.038748081773519516,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.2636070251464844,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.039233580231666565,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.2751896381378174,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.03429320082068443,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.2765679359436035,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.03267038241028786,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.2387681007385254,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.03349406272172928,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.282162666320801,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.032002877444028854,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.246623992919922,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.031509071588516235,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.2374932765960693,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.02958357334136963,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.24362850189209,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.03248635679483414,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.237210273742676,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.03230655938386917,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.251129150390625,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.029688173905014992,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.2141575813293457,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.027773940935730934,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.264249086380005,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.02687031403183937,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.2513632774353027,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.02503623627126217,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.217961072921753,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.027071228250861168,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.2784645557403564,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.02600773423910141,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.2641382217407227,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.026998162269592285,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.2646782398223877,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.025579402223229408,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.2700581550598145,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.02645333670079708,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.2545266151428223,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.02717350795865059,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.2795705795288086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.02675914391875267,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.2936904430389404,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.02751768007874489,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.2329440116882324,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.025446882471442223,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.274294376373291,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.02365768514573574,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.2585229873657227,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.0266884732991457,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.2872962951660156,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.02572469227015972,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.253828287124634,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.02540273219347,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.303926944732666,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.024694625288248062,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.264145851135254,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.027233170345425606,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.286924362182617,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02635285072028637,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.2930331230163574,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.025157805532217026,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.2906906604766846,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.024729596450924873,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.277128219604492,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.023809000849723816,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.2957217693328857,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02474421076476574,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.27791690826416,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.02439294196665287,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.2912981510162354,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.023601334542036057,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.317446231842041,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.0253264419734478,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.2742302417755127,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.02389134280383587,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.281032085418701,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.02451813407242298,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.3339295387268066,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.024324845522642136,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.290402889251709,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.024939008057117462,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.2933621406555176,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.024450505152344704,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.3055014610290527,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.02262270078063011,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.2964205741882324,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.02311946265399456,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.3102402687072754,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.024362705647945404,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.3020925521850586,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.024785032495856285,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.287290096282959,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.02358846180140972,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.334261894226074,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.024330584332346916,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.2799081802368164,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.023546984419226646,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.3228847980499268,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.023060832172632217,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.325793981552124,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.02388455718755722,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.2892062664031982,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.023740701377391815,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.2969369888305664,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.02458217367529869,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.315610408782959,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.023173579946160316,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.2561936378479004,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.024137936532497406,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.2839202880859375,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.022436272352933884,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.3057351112365723,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.023997878655791283,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.3583154678344727,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.024119896814227104,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.322892427444458,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.022283365949988365,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.3145735263824463,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.023433363065123558,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.3564529418945312,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.02224746160209179,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.2985916137695312,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.023786570876836777,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.302539587020874,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.025200961157679558,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.309976577758789,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.023849809542298317,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.343660354614258,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.023311031982302666,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.2978909015655518,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.023174891248345375,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.326188802719116,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.02681034244596958,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 3.3422017097473145,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.109009265899658,
+ "eval_runtime": 58.0506,
+ "eval_samples_per_second": 42.067,
+ "eval_steps_per_second": 1.326,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.028057295829057693,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.1825037002563477,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.034786805510520935,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.200389862060547,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.03751441463828087,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.2047512531280518,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.035590436309576035,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.1735124588012695,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.03426426649093628,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.199483871459961,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.037030890583992004,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.213322639465332,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.032468877732753754,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.179354190826416,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.027202758938074112,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.216036319732666,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.031184563413262367,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.208205223083496,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.029414812102913857,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.180370569229126,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.02679901383817196,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.210325002670288,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.02863631211221218,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.1601076126098633,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.029135366901755333,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.2262887954711914,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.030535712838172913,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.23160719871521,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.030563877895474434,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.255211353302002,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.02906505949795246,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.2006797790527344,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.030871668830513954,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.2442867755889893,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.02913421392440796,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.218442440032959,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.030087953433394432,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.2164535522460938,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.034147944301366806,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.218660831451416,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.03593966364860535,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.2240614891052246,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.032090652734041214,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.2309441566467285,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.02776424027979374,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.2702016830444336,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.027174271643161774,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.19810152053833,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.028479408472776413,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.2178421020507812,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.026820644736289978,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.2634482383728027,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.026427917182445526,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.2452914714813232,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.02679610252380371,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.243825912475586,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.026115626096725464,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.218787670135498,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.026686571538448334,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.2785303592681885,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.02685859426856041,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.2572226524353027,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.026968171820044518,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.2625553607940674,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.025397246703505516,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.262303352355957,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.025660471990704536,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.269401788711548,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.02713891491293907,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.2416529655456543,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.027591092512011528,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.2814316749572754,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.024462848901748657,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.249725103378296,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.024868478998541832,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.231574535369873,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.027015401050448418,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.282627582550049,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.02678140625357628,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.2208900451660156,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.026830527931451797,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.260575294494629,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.02430890128016472,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.2522997856140137,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.027047613635659218,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.2721078395843506,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.026861369609832764,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.2235946655273438,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.026650549843907356,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.302743434906006,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.02551819011569023,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.265763282775879,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.024843450635671616,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.272397518157959,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.024423206225037575,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.253495454788208,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.024180661886930466,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.260438919067383,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.023554982617497444,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.2342851161956787,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.023470086976885796,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.251622438430786,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.024405231699347496,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.2588815689086914,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.024147814139723778,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.2723565101623535,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.023217910900712013,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.2555792331695557,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.024921158328652382,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.26884126663208,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.02455919235944748,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.2674927711486816,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.024725932627916336,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.2511110305786133,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.025769414380192757,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.290651321411133,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.025928806513547897,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.302396297454834,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.02499072253704071,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.2813124656677246,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.025431277230381966,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.234684705734253,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.026871860027313232,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.2800331115722656,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.02578776702284813,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.3117146492004395,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.026318378746509552,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.257615089416504,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.025542527437210083,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.2969868183135986,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.025723429396748543,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.290259838104248,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.024588774889707565,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.284090280532837,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.024619122967123985,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.284879207611084,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.024441618472337723,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.283104658126831,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.02402760274708271,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.3013617992401123,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.025028521195054054,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.2892568111419678,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.029445001855492592,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 3.2474446296691895,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.128176689147949,
+ "eval_runtime": 58.8588,
+ "eval_samples_per_second": 41.489,
+ "eval_steps_per_second": 1.308,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.0298826452344656,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.138103485107422,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.0387548953294754,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.161492347717285,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.03816070780158043,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.1739673614501953,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.03409738838672638,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.182478189468384,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.033135563135147095,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.1720199584960938,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.03203624114394188,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.1699862480163574,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.031237754970788956,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.179363250732422,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.03342146426439285,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.1822280883789062,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.03561089560389519,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.1767094135284424,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.034710463136434555,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.218817949295044,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.030870331451296806,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.168602705001831,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.028004391118884087,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.224155902862549,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.030953701585531235,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.1789557933807373,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.030289432033896446,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.1890387535095215,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.030684132128953934,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.195427417755127,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.03022196516394615,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.1984965801239014,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.026935536414384842,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.2014272212982178,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.02702922560274601,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.205538034439087,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.026682600378990173,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.1954591274261475,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.0273384228348732,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.1782338619232178,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.028587577864527702,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.199565887451172,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.028848132118582726,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.181786298751831,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.027535341680049896,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.2013628482818604,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.02756025455892086,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.1749587059020996,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.027358995750546455,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.161881923675537,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.025918981060385704,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.225224256515503,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.026479976251721382,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.2129316329956055,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.02685079537332058,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.1651554107666016,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.027378706261515617,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.193108558654785,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.027240974828600883,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.190605640411377,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.028065402060747147,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.2068259716033936,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.028329014778137207,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.2316060066223145,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.028920065611600876,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.1933932304382324,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.028299354016780853,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.192711353302002,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.025996020063757896,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.243171453475952,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.0264595914632082,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.199758529663086,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.02704528719186783,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.2390713691711426,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.026571933180093765,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.219440460205078,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.025562983006238937,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.2538228034973145,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.024683184921741486,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.2372591495513916,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.025656528770923615,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.250558853149414,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.024807540699839592,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.1856563091278076,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.026149194687604904,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.172186851501465,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.02617705799639225,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.250445604324341,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.026542512699961662,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.220273494720459,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.024963509291410446,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.2330946922302246,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.027144264429807663,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.1860971450805664,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.02665683813393116,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.257209062576294,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.024872658774256706,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.2266182899475098,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.025235185399651527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.2134504318237305,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.025029495358467102,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.2150087356567383,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.025630896911025047,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.282879114151001,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.02722341939806938,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.2366206645965576,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.02522372268140316,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.2840523719787598,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.02494790591299534,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.2405319213867188,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.026662737131118774,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.239692211151123,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.026082608848810196,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.2156195640563965,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.02423974685370922,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.1863760948181152,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.025226688012480736,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.20874285697937,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.02428225241601467,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.2209601402282715,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.023819707334041595,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.2564258575439453,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.025661710649728775,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.244133949279785,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.02423049882054329,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.2376630306243896,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.02382473088800907,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.2951290607452393,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.02488875202834606,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.2848310470581055,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.02575223706662655,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.2574503421783447,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.02462848648428917,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.2464442253112793,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.024912388995289803,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.2459192276000977,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.025020884349942207,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.213622808456421,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.026330886408686638,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.2305431365966797,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.02708602324128151,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.2487688064575195,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.030908891931176186,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 3.2110159397125244,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.146673202514648,
+ "eval_runtime": 58.7379,
+ "eval_samples_per_second": 41.574,
+ "eval_steps_per_second": 1.311,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.02919941022992134,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.1533267498016357,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.03379782289266586,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.1436538696289062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.033698830753564835,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.15297532081604,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.03077736124396324,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.128848075866699,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.031112324446439743,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.1461119651794434,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.03346852958202362,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.113943099975586,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.03307225927710533,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.147498607635498,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.030230754986405373,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.1162924766540527,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.02982981689274311,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.144481658935547,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.031735632568597794,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.133862018585205,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.0312766470015049,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.110287666320801,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.029902562499046326,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.1384904384613037,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.0306564774364233,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.1138200759887695,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.03084711916744709,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.155240535736084,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.03183220326900482,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.1593515872955322,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.031826410442590714,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.1517653465270996,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.030141707509756088,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.1291589736938477,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.03158571943640709,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.178101062774658,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.03510458394885063,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.138767719268799,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.03660859912633896,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.1708762645721436,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.034004807472229004,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.1404073238372803,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.03077496401965618,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.159081220626831,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.030384650453925133,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.1872847080230713,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.02957277186214924,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.165973424911499,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.029529055580496788,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.123840808868408,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.027776138857007027,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.1705269813537598,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.028943249955773354,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.1181654930114746,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.028178401291370392,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.201678991317749,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.02865665778517723,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.18479061126709,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.029173649847507477,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.1844916343688965,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.0280170775949955,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.187105417251587,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.029103878885507584,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.183281660079956,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.029045486822724342,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.1848196983337402,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.028364380821585655,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.197348117828369,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.029651399701833725,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.1429920196533203,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.0281438659876585,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.1916513442993164,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.029437677934765816,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.172391176223755,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.03433044254779816,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.2296864986419678,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.03259492665529251,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.1962363719940186,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.0299630519002676,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.169673442840576,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.028495769947767258,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.1855409145355225,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.0290420800447464,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.23667573928833,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.03340664133429527,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.1873788833618164,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.03253626823425293,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.204796552658081,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.028142794966697693,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.193922281265259,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.028805622830986977,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.1863901615142822,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.030033612623810768,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.198482036590576,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.028154220432043076,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.215069055557251,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.028098132461309433,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.2197628021240234,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.028679659590125084,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.2062196731567383,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.029104258865118027,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.2182908058166504,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.029629778116941452,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.2165639400482178,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.028304366394877434,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.1783785820007324,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.026900488883256912,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.1876602172851562,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.027047287672758102,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.1879465579986572,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.027424108237028122,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.2407851219177246,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.029319508001208305,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.1980130672454834,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.027114372700452805,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.211357593536377,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.02579840086400509,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.2067861557006836,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.02767331339418888,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.198587417602539,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.026645265519618988,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.177441120147705,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.026181906461715698,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.226815700531006,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.027218692004680634,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.1731510162353516,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.027551919221878052,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.2268621921539307,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.025782009586691856,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.2052621841430664,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.024231620132923126,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.201770544052124,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.02474607340991497,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.1935834884643555,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.026299327611923218,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.223275899887085,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.02439914271235466,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.2513954639434814,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.02650653012096882,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.1914851665496826,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.024847298860549927,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.2173104286193848,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.028475770726799965,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 3.2184863090515137,
+ "step": 2160
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.208840083996672e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2160/training_args.bin b/runs/l2r90-baseline/checkpoint-2160/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2160/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-2520/config.json b/runs/l2r90-baseline/checkpoint-2520/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-2520/generation_config.json b/runs/l2r90-baseline/checkpoint-2520/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-2520/model.safetensors b/runs/l2r90-baseline/checkpoint-2520/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9a395ee4cf04af48594b64d3a1216b531f6ad6ea
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9a2d0d096c5824a8aa66c26ef12ef67dac462778badde06e4dbad522cf179a19
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-2520/optimizer.pt b/runs/l2r90-baseline/checkpoint-2520/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d749555ccbeb8066775e6261e820ab2f8596307b
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a0b2388cb0aa77011ebb6d03d323ee12143d66821fc79bb665fa4113faf9f05c
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..eaeeea885c2ef4b0cef75101a8e62f4ab4851bd2
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d1e03ae1cb1232d1d92f456eeb7fed9b86dbea40002020bd21a8bff940afba04
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..81543e2df97bda93c650637e0e76e6eba567cc34
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:204f99921454e515b3056a84d1cf89d690447610764df571b00287728068baff
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4ff4a5c1c578c912bb73e9e1039feddae85108d0
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9add74f0deb40ada69213c31c9eb2148f496716039a30fb746923520918c78da
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fc461818ec193e0f1784351202790a5fa22189ee
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ff9312b009886e7daf125d940b7a052dd2e1a5c49cb82b56a0aaa5ca1c30fb9a
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2520/scheduler.pt b/runs/l2r90-baseline/checkpoint-2520/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b53f13f877641b522f2c435934f587fdad6161cc
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6f776ea81adcefd29eff4d7371e55373e8890bd750697e1c8b781d131e0a39a6
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-2520/tokenizer.json b/runs/l2r90-baseline/checkpoint-2520/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2520/trainer_state.json b/runs/l2r90-baseline/checkpoint-2520/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3a1defa4787fc5278ed37bc19202dcd5bdc21204
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/trainer_state.json
@@ -0,0 +1,17946 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 35.0,
+ "eval_steps": 500,
+ "global_step": 2520,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.078423500061035,
+ "eval_runtime": 57.3189,
+ "eval_samples_per_second": 42.604,
+ "eval_steps_per_second": 1.343,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.02624760940670967,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.261288642883301,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.03552987053990364,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.2682106494903564,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.040380869060754776,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.2813782691955566,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.03861293941736221,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.258193016052246,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.04091731086373329,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.3052220344543457,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.043269723653793335,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.2820217609405518,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.04580320790410042,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.312005043029785,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.047336459159851074,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.3217713832855225,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.04587360844016075,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.2869949340820312,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.04002532362937927,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.2879223823547363,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.0368693508207798,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.2983577251434326,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.037317775189876556,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.3062524795532227,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.03655700013041496,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.31072735786438,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.033466216176748276,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.2954514026641846,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.030785970389842987,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.27952241897583,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.029669342562556267,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.2613229751586914,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.033505845814943314,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.2987258434295654,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.032395437359809875,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.296708583831787,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.031322430819272995,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.3160252571105957,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.029345186427235603,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.307143211364746,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.026715675368905067,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.286012649536133,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.028592750430107117,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.2528645992279053,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.028861764818429947,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.2795748710632324,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.02609528973698616,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.326317071914673,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.027718454599380493,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.268110513687134,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.026382867246866226,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.294499397277832,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.027083583176136017,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.3202130794525146,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.02516467310488224,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.3485021591186523,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.024148069322109222,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.3099541664123535,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.024552857503294945,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.2885942459106445,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.02551114559173584,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.3232040405273438,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.025423647835850716,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.3097832202911377,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.023211417719721794,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.3075966835021973,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.024623898789286613,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.2917377948760986,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.02677663415670395,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.356265068054199,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.02673642337322235,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.314903736114502,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.024466225877404213,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.30912446975708,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.025370748713612556,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.363640069961548,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.026536114513874054,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.304851531982422,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.026536893099546432,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.342902898788452,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.02645307406783104,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.3164432048797607,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.02523060515522957,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.34376859664917,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.02590431459248066,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.3479061126708984,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.0230655986815691,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.2890982627868652,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.02403893694281578,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.345456600189209,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.02238907851278782,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.3471767902374268,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.021377569064497948,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.319232225418091,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.02272353693842888,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.316892623901367,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.02413824200630188,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.3450968265533447,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.02334682270884514,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.3640642166137695,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.021966520696878433,
+ "learning_rate": 0.000303,
+ "loss": 3.358555316925049,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.023984894156455994,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.358398914337158,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.023520268499851227,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.3224587440490723,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.0231285709887743,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.3671562671661377,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.022818170487880707,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.361027717590332,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.02280694991350174,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.2923383712768555,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.021362345665693283,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.3346734046936035,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.02372218668460846,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.354748249053955,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.022949764505028725,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.3857789039611816,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.021954255178570747,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.343330144882202,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.022438595071434975,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.328618049621582,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.021969925612211227,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.355971574783325,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.022706981748342514,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.3529577255249023,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.024083172902464867,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.348170757293701,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.023259354755282402,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.3474369049072266,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.02218157798051834,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.369053840637207,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.02305331453680992,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.3622641563415527,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.022796178236603737,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.351510524749756,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.021299205720424652,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.339406728744507,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.022786924615502357,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.3182196617126465,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.021729128435254097,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.383904218673706,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.024604283273220062,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 3.2912864685058594,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.0894880294799805,
+ "eval_runtime": 58.3967,
+ "eval_samples_per_second": 41.817,
+ "eval_steps_per_second": 1.319,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.02673362009227276,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.270986795425415,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.04036599025130272,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.2358999252319336,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.047250088304281235,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.2706189155578613,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.04642819985747337,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.236340284347534,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.039405327290296555,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.203136682510376,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.038748081773519516,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.2636070251464844,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.039233580231666565,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.2751896381378174,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.03429320082068443,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.2765679359436035,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.03267038241028786,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.2387681007385254,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.03349406272172928,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.282162666320801,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.032002877444028854,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.246623992919922,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.031509071588516235,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.2374932765960693,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.02958357334136963,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.24362850189209,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.03248635679483414,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.237210273742676,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.03230655938386917,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.251129150390625,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.029688173905014992,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.2141575813293457,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.027773940935730934,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.264249086380005,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.02687031403183937,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.2513632774353027,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.02503623627126217,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.217961072921753,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.027071228250861168,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.2784645557403564,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.02600773423910141,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.2641382217407227,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.026998162269592285,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.2646782398223877,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.025579402223229408,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.2700581550598145,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.02645333670079708,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.2545266151428223,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.02717350795865059,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.2795705795288086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.02675914391875267,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.2936904430389404,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.02751768007874489,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.2329440116882324,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.025446882471442223,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.274294376373291,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.02365768514573574,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.2585229873657227,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.0266884732991457,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.2872962951660156,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.02572469227015972,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.253828287124634,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.02540273219347,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.303926944732666,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.024694625288248062,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.264145851135254,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.027233170345425606,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.286924362182617,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02635285072028637,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.2930331230163574,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.025157805532217026,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.2906906604766846,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.024729596450924873,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.277128219604492,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.023809000849723816,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.2957217693328857,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02474421076476574,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.27791690826416,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.02439294196665287,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.2912981510162354,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.023601334542036057,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.317446231842041,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.0253264419734478,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.2742302417755127,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.02389134280383587,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.281032085418701,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.02451813407242298,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.3339295387268066,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.024324845522642136,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.290402889251709,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.024939008057117462,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.2933621406555176,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.024450505152344704,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.3055014610290527,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.02262270078063011,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.2964205741882324,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.02311946265399456,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.3102402687072754,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.024362705647945404,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.3020925521850586,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.024785032495856285,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.287290096282959,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.02358846180140972,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.334261894226074,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.024330584332346916,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.2799081802368164,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.023546984419226646,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.3228847980499268,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.023060832172632217,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.325793981552124,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.02388455718755722,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.2892062664031982,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.023740701377391815,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.2969369888305664,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.02458217367529869,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.315610408782959,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.023173579946160316,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.2561936378479004,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.024137936532497406,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.2839202880859375,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.022436272352933884,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.3057351112365723,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.023997878655791283,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.3583154678344727,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.024119896814227104,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.322892427444458,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.022283365949988365,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.3145735263824463,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.023433363065123558,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.3564529418945312,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.02224746160209179,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.2985916137695312,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.023786570876836777,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.302539587020874,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.025200961157679558,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.309976577758789,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.023849809542298317,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.343660354614258,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.023311031982302666,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.2978909015655518,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.023174891248345375,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.326188802719116,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.02681034244596958,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 3.3422017097473145,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.109009265899658,
+ "eval_runtime": 58.0506,
+ "eval_samples_per_second": 42.067,
+ "eval_steps_per_second": 1.326,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.028057295829057693,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.1825037002563477,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.034786805510520935,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.200389862060547,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.03751441463828087,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.2047512531280518,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.035590436309576035,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.1735124588012695,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.03426426649093628,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.199483871459961,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.037030890583992004,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.213322639465332,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.032468877732753754,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.179354190826416,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.027202758938074112,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.216036319732666,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.031184563413262367,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.208205223083496,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.029414812102913857,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.180370569229126,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.02679901383817196,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.210325002670288,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.02863631211221218,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.1601076126098633,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.029135366901755333,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.2262887954711914,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.030535712838172913,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.23160719871521,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.030563877895474434,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.255211353302002,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.02906505949795246,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.2006797790527344,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.030871668830513954,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.2442867755889893,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.02913421392440796,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.218442440032959,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.030087953433394432,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.2164535522460938,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.034147944301366806,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.218660831451416,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.03593966364860535,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.2240614891052246,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.032090652734041214,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.2309441566467285,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.02776424027979374,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.2702016830444336,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.027174271643161774,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.19810152053833,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.028479408472776413,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.2178421020507812,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.026820644736289978,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.2634482383728027,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.026427917182445526,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.2452914714813232,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.02679610252380371,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.243825912475586,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.026115626096725464,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.218787670135498,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.026686571538448334,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.2785303592681885,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.02685859426856041,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.2572226524353027,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.026968171820044518,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.2625553607940674,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.025397246703505516,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.262303352355957,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.025660471990704536,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.269401788711548,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.02713891491293907,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.2416529655456543,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.027591092512011528,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.2814316749572754,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.024462848901748657,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.249725103378296,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.024868478998541832,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.231574535369873,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.027015401050448418,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.282627582550049,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.02678140625357628,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.2208900451660156,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.026830527931451797,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.260575294494629,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.02430890128016472,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.2522997856140137,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.027047613635659218,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.2721078395843506,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.026861369609832764,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.2235946655273438,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.026650549843907356,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.302743434906006,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.02551819011569023,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.265763282775879,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.024843450635671616,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.272397518157959,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.024423206225037575,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.253495454788208,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.024180661886930466,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.260438919067383,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.023554982617497444,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.2342851161956787,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.023470086976885796,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.251622438430786,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.024405231699347496,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.2588815689086914,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.024147814139723778,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.2723565101623535,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.023217910900712013,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.2555792331695557,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.024921158328652382,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.26884126663208,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.02455919235944748,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.2674927711486816,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.024725932627916336,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.2511110305786133,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.025769414380192757,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.290651321411133,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.025928806513547897,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.302396297454834,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.02499072253704071,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.2813124656677246,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.025431277230381966,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.234684705734253,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.026871860027313232,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.2800331115722656,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.02578776702284813,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.3117146492004395,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.026318378746509552,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.257615089416504,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.025542527437210083,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.2969868183135986,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.025723429396748543,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.290259838104248,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.024588774889707565,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.284090280532837,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.024619122967123985,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.284879207611084,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.024441618472337723,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.283104658126831,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.02402760274708271,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.3013617992401123,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.025028521195054054,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.2892568111419678,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.029445001855492592,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 3.2474446296691895,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.128176689147949,
+ "eval_runtime": 58.8588,
+ "eval_samples_per_second": 41.489,
+ "eval_steps_per_second": 1.308,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.0298826452344656,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.138103485107422,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.0387548953294754,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.161492347717285,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.03816070780158043,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.1739673614501953,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.03409738838672638,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.182478189468384,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.033135563135147095,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.1720199584960938,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.03203624114394188,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.1699862480163574,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.031237754970788956,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.179363250732422,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.03342146426439285,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.1822280883789062,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.03561089560389519,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.1767094135284424,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.034710463136434555,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.218817949295044,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.030870331451296806,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.168602705001831,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.028004391118884087,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.224155902862549,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.030953701585531235,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.1789557933807373,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.030289432033896446,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.1890387535095215,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.030684132128953934,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.195427417755127,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.03022196516394615,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.1984965801239014,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.026935536414384842,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.2014272212982178,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.02702922560274601,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.205538034439087,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.026682600378990173,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.1954591274261475,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.0273384228348732,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.1782338619232178,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.028587577864527702,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.199565887451172,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.028848132118582726,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.181786298751831,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.027535341680049896,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.2013628482818604,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.02756025455892086,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.1749587059020996,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.027358995750546455,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.161881923675537,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.025918981060385704,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.225224256515503,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.026479976251721382,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.2129316329956055,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.02685079537332058,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.1651554107666016,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.027378706261515617,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.193108558654785,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.027240974828600883,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.190605640411377,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.028065402060747147,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.2068259716033936,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.028329014778137207,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.2316060066223145,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.028920065611600876,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.1933932304382324,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.028299354016780853,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.192711353302002,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.025996020063757896,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.243171453475952,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.0264595914632082,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.199758529663086,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.02704528719186783,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.2390713691711426,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.026571933180093765,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.219440460205078,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.025562983006238937,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.2538228034973145,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.024683184921741486,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.2372591495513916,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.025656528770923615,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.250558853149414,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.024807540699839592,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.1856563091278076,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.026149194687604904,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.172186851501465,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.02617705799639225,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.250445604324341,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.026542512699961662,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.220273494720459,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.024963509291410446,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.2330946922302246,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.027144264429807663,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.1860971450805664,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.02665683813393116,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.257209062576294,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.024872658774256706,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.2266182899475098,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.025235185399651527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.2134504318237305,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.025029495358467102,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.2150087356567383,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.025630896911025047,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.282879114151001,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.02722341939806938,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.2366206645965576,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.02522372268140316,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.2840523719787598,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.02494790591299534,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.2405319213867188,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.026662737131118774,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.239692211151123,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.026082608848810196,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.2156195640563965,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.02423974685370922,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.1863760948181152,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.025226688012480736,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.20874285697937,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.02428225241601467,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.2209601402282715,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.023819707334041595,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.2564258575439453,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.025661710649728775,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.244133949279785,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.02423049882054329,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.2376630306243896,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.02382473088800907,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.2951290607452393,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.02488875202834606,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.2848310470581055,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.02575223706662655,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.2574503421783447,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.02462848648428917,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.2464442253112793,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.024912388995289803,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.2459192276000977,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.025020884349942207,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.213622808456421,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.026330886408686638,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.2305431365966797,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.02708602324128151,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.2487688064575195,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.030908891931176186,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 3.2110159397125244,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.146673202514648,
+ "eval_runtime": 58.7379,
+ "eval_samples_per_second": 41.574,
+ "eval_steps_per_second": 1.311,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.02919941022992134,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.1533267498016357,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.03379782289266586,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.1436538696289062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.033698830753564835,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.15297532081604,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.03077736124396324,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.128848075866699,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.031112324446439743,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.1461119651794434,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.03346852958202362,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.113943099975586,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.03307225927710533,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.147498607635498,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.030230754986405373,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.1162924766540527,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.02982981689274311,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.144481658935547,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.031735632568597794,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.133862018585205,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.0312766470015049,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.110287666320801,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.029902562499046326,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.1384904384613037,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.0306564774364233,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.1138200759887695,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.03084711916744709,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.155240535736084,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.03183220326900482,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.1593515872955322,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.031826410442590714,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.1517653465270996,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.030141707509756088,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.1291589736938477,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.03158571943640709,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.178101062774658,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.03510458394885063,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.138767719268799,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.03660859912633896,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.1708762645721436,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.034004807472229004,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.1404073238372803,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.03077496401965618,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.159081220626831,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.030384650453925133,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.1872847080230713,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.02957277186214924,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.165973424911499,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.029529055580496788,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.123840808868408,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.027776138857007027,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.1705269813537598,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.028943249955773354,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.1181654930114746,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.028178401291370392,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.201678991317749,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.02865665778517723,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.18479061126709,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.029173649847507477,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.1844916343688965,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.0280170775949955,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.187105417251587,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.029103878885507584,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.183281660079956,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.029045486822724342,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.1848196983337402,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.028364380821585655,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.197348117828369,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.029651399701833725,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.1429920196533203,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.0281438659876585,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.1916513442993164,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.029437677934765816,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.172391176223755,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.03433044254779816,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.2296864986419678,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.03259492665529251,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.1962363719940186,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.0299630519002676,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.169673442840576,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.028495769947767258,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.1855409145355225,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.0290420800447464,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.23667573928833,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.03340664133429527,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.1873788833618164,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.03253626823425293,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.204796552658081,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.028142794966697693,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.193922281265259,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.028805622830986977,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.1863901615142822,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.030033612623810768,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.198482036590576,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.028154220432043076,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.215069055557251,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.028098132461309433,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.2197628021240234,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.028679659590125084,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.2062196731567383,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.029104258865118027,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.2182908058166504,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.029629778116941452,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.2165639400482178,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.028304366394877434,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.1783785820007324,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.026900488883256912,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.1876602172851562,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.027047287672758102,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.1879465579986572,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.027424108237028122,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.2407851219177246,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.029319508001208305,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.1980130672454834,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.027114372700452805,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.211357593536377,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.02579840086400509,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.2067861557006836,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.02767331339418888,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.198587417602539,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.026645265519618988,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.177441120147705,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.026181906461715698,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.226815700531006,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.027218692004680634,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.1731510162353516,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.027551919221878052,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.2268621921539307,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.025782009586691856,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.2052621841430664,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.024231620132923126,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.201770544052124,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.02474607340991497,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.1935834884643555,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.026299327611923218,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.223275899887085,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.02439914271235466,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.2513954639434814,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.02650653012096882,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.1914851665496826,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.024847298860549927,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.2173104286193848,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.028475770726799965,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 3.2184863090515137,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.168707847595215,
+ "eval_runtime": 57.4035,
+ "eval_samples_per_second": 42.541,
+ "eval_steps_per_second": 1.341,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.02837727963924408,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.119239330291748,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.03562312573194504,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.092813014984131,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.0344194732606411,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.0802855491638184,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.03188958019018173,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.1177961826324463,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.032181013375520706,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.1343281269073486,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.031085725873708725,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.118828296661377,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.031995322555303574,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.110109329223633,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.02949182130396366,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.0972743034362793,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.029299668967723846,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.1257238388061523,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.030558617785573006,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.1216022968292236,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.03181575983762741,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.093721866607666,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.030934303998947144,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.0910396575927734,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.030235569924116135,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.133371114730835,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.03350163996219635,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.1178882122039795,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.03468719869852066,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.1222469806671143,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.034203238785266876,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.12453031539917,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.03321516141295433,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.115428924560547,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.03298434242606163,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.127288341522217,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.029987169429659843,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.1194515228271484,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.031271349638700485,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.1284165382385254,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.032201699912548065,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.1912736892700195,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.0299740731716156,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.1233785152435303,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.029613154008984566,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.1535067558288574,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.03303765505552292,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.1550910472869873,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.032644059509038925,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.1367573738098145,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.031695179641246796,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.1350858211517334,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.030685821548104286,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.121593952178955,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.03091946430504322,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.113060235977173,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.0282905176281929,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.184551239013672,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.030020160600543022,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.1673004627227783,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.031114062294363976,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.162588596343994,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.03007666766643524,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.132836103439331,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.029013793915510178,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.1552720069885254,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.02784561738371849,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.189272165298462,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.02759280428290367,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.141814708709717,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.02699868381023407,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.1127257347106934,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.02782190404832363,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.193972587585449,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.027298318222165108,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.1393446922302246,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.027321161702275276,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.0926666259765625,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.026399722322821617,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.1578259468078613,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.02770969271659851,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.1816914081573486,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.028344258666038513,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.1860158443450928,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.028091782703995705,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.1585092544555664,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.026564039289951324,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.131692409515381,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.026349127292633057,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.160431385040283,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.030081138014793396,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.1647543907165527,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.029250919818878174,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.1522226333618164,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.0282477755099535,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.139249324798584,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.0284534003585577,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.154621124267578,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.02677394263446331,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.1375081539154053,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.026511570438742638,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.1355676651000977,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.02766270563006401,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.143874168395996,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.027013977989554405,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.1602487564086914,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.026629449799656868,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.1549456119537354,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.0272684209048748,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.1739020347595215,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.027645915746688843,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.172959327697754,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.025413552299141884,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.1649041175842285,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.028406929224729538,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.1821887493133545,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.027753977105021477,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.163238048553467,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.02693209983408451,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.1921019554138184,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.02671878971159458,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.159438133239746,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.027099428698420525,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.1806793212890625,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.025098824873566628,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.158310651779175,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.02655479498207569,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.1927309036254883,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.025836164131760597,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.1578540802001953,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.02657160349190235,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.1658592224121094,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.027223270386457443,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.1760241985321045,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.025695348158478737,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.202389717102051,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.025819813832640648,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.2111103534698486,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.02621106244623661,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.1560299396514893,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.026585502550005913,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.176602840423584,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.03038652054965496,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 3.1799943447113037,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.193845748901367,
+ "eval_runtime": 57.9911,
+ "eval_samples_per_second": 42.11,
+ "eval_steps_per_second": 1.328,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.03135736659169197,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.070986270904541,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.038986921310424805,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.0708835124969482,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.035115309059619904,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.1070001125335693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.032868221402168274,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.082258701324463,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.03703184798359871,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.057744264602661,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.03831101953983307,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.0789475440979004,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.036301176995038986,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.0565457344055176,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.03166409209370613,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.0574910640716553,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.02984458953142166,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.080018997192383,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.034001126885414124,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.0813302993774414,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.03814122825860977,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.0579185485839844,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.03870106488466263,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.0857529640197754,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.04005642607808113,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.0880727767944336,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.03760755434632301,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.0795767307281494,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.03260907158255577,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.1006722450256348,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.029866183176636696,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.0850343704223633,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.030988436192274094,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.0743892192840576,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.03698134422302246,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.056002378463745,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.03696191683411598,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.069124698638916,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.03453022986650467,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.1017374992370605,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.032458193600177765,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.1166419982910156,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.02892368845641613,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.0863287448883057,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.030625877901911736,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.0938801765441895,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.030280394479632378,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.086432933807373,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.029181072488427162,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.0977859497070312,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.028039127588272095,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.112367868423462,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.02910272404551506,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.1253533363342285,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.03086872771382332,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.103689670562744,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.02883610874414444,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.111599922180176,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.028490116819739342,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.1410303115844727,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.028714407235383987,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.133023738861084,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.029553238302469254,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.107003688812256,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.02997349575161934,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.135680675506592,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.029087591916322708,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.1103882789611816,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.029947761446237564,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.0961623191833496,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.02950555644929409,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.1162495613098145,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.029569562524557114,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.096372127532959,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.031366728246212006,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.120089530944824,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.0313473716378212,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.100785732269287,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.028021851554512978,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.0940558910369873,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.027477117255330086,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.1210484504699707,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.029690749943256378,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.1350059509277344,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.029659733176231384,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.1220788955688477,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.028260033577680588,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.1000189781188965,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.027554461732506752,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.1044349670410156,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.027929967269301414,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.1353745460510254,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.029309319332242012,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.1578292846679688,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.029183287173509598,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.1217150688171387,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.02707512304186821,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.119617223739624,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.03006536327302456,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.1481480598449707,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.03026784397661686,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.142770767211914,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.02827967144548893,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.1582541465759277,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.028892774134874344,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.1530709266662598,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.03023783676326275,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.1102521419525146,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.027951840311288834,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.139113426208496,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.02807226963341236,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.149428367614746,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.029721451923251152,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.120518207550049,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.02867680788040161,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.1109931468963623,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.02663845382630825,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.116053581237793,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.028366081416606903,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.1444268226623535,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.028931576758623123,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.163870096206665,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.02664567157626152,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.1349146366119385,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.028031354770064354,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.1638219356536865,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.029421644285321236,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.1601340770721436,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.027983693405985832,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.1636788845062256,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.02719871513545513,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.1246869564056396,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.03030397929251194,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.1749579906463623,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.029340893030166626,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.1751790046691895,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.027624106034636497,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.178704023361206,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.02830546908080578,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.177117347717285,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.03168068826198578,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.125427722930908,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.033327918499708176,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 3.13071870803833,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.218388080596924,
+ "eval_runtime": 57.611,
+ "eval_samples_per_second": 42.388,
+ "eval_steps_per_second": 1.337,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.03336749225854874,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.064338445663452,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.038479577749967575,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.0478687286376953,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.032562077045440674,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.019803762435913,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.0331728458404541,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.0599141120910645,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.03361169621348381,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.0315823554992676,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.032973092049360275,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.025928258895874,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.030388766899704933,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.070671558380127,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.030899515375494957,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.012938976287842,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.0335412323474884,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.0156331062316895,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.035731345415115356,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.0578935146331787,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.03379053995013237,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.048701286315918,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.031242679804563522,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.0832977294921875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.030997611582279205,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.0663583278656006,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.03397331386804581,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.0115303993225098,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.03349050134420395,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.02864933013916,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.031079065054655075,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.053401470184326,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.029346197843551636,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.0700926780700684,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.030784208327531815,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.066525936126709,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.03368140384554863,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.0956242084503174,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.03356792405247688,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.0394320487976074,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.029909325763583183,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.031525135040283,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.030711578205227852,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.075474977493286,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.03452072665095329,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.0228147506713867,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.03235075995326042,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.0664501190185547,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.029671598225831985,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.055302143096924,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.033868469297885895,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.053016185760498,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.0329158715903759,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.1017332077026367,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.031983695924282074,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.0706472396850586,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.02963697351515293,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.0711426734924316,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.03162391856312752,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.044867515563965,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.030773654580116272,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.0963196754455566,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.031220417469739914,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.0892245769500732,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.030928561463952065,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.1181323528289795,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.03192073106765747,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.066584348678589,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.03075495734810829,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.0436511039733887,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.02895325981080532,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.123373031616211,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.029738714918494225,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.1049060821533203,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.031937118619680405,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.099111318588257,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.029886579141020775,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.0970301628112793,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.029569782316684723,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.062499761581421,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.02924283780157566,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.0618538856506348,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.030955800786614418,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.0745606422424316,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.03145027533173561,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.122938632965088,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.028926467522978783,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.0808773040771484,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.029592076316475868,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.1167049407958984,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.028070243075489998,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.0781636238098145,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.028642063960433006,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.093937397003174,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.0290757454931736,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.1079559326171875,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.027867283672094345,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.0885753631591797,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.02896495908498764,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.0889246463775635,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.02777806483209133,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.084439277648926,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.02904277853667736,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.125246047973633,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.029010627418756485,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.111422061920166,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.028178894892334938,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.115067481994629,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.02763799950480461,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.1228506565093994,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.028052521869540215,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.1150546073913574,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.029438253492116928,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.0921788215637207,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.028195543214678764,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.1408450603485107,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.02937602624297142,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.111612319946289,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.028430236503481865,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.093441963195801,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.027624569833278656,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.1184372901916504,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.029797039926052094,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.1074559688568115,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.028728218749165535,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.161461114883423,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.02704412117600441,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.13325834274292,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.028067929670214653,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.093172550201416,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.028848597779870033,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.1461825370788574,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.02956785447895527,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.141045570373535,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.028214773163199425,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.136033773422241,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.029155246913433075,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.0986382961273193,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.02968468703329563,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.11630916595459,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.028356682509183884,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.1134822368621826,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.03301718086004257,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 3.118591785430908,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.246211528778076,
+ "eval_runtime": 57.5155,
+ "eval_samples_per_second": 42.458,
+ "eval_steps_per_second": 1.339,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.033637888729572296,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.0192506313323975,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.04057888314127922,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 2.977440595626831,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.03579854592680931,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 2.9902219772338867,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.03131850063800812,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 2.9988656044006348,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.03262636810541153,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.037747621536255,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.03260403499007225,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 2.987508773803711,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.033165134489536285,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.0511183738708496,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.03303489834070206,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.0083041191101074,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.03185006603598595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 2.9966228008270264,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.03135651350021362,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.036759376525879,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.03216925635933876,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.0304150581359863,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.03231671452522278,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.0300474166870117,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.03263348713517189,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.0298163890838623,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.032594650983810425,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 2.98789119720459,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.03406650573015213,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.0102760791778564,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.03379780799150467,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.0564732551574707,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.03172587603330612,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.0410943031311035,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.02938298135995865,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.0517544746398926,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.03206252679228783,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.0352776050567627,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.033386752009391785,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.020535945892334,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.033984702080488205,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.0552730560302734,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0329931266605854,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.0075573921203613,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.032101184129714966,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.046388864517212,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.032873060554265976,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.0340287685394287,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.030672110617160797,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.0347838401794434,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.030014054849743843,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.0315332412719727,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.02932834066450596,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.084340810775757,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.029181934893131256,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.032886028289795,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.02966221794486046,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.032766819000244,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.029713116586208344,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.028778076171875,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.029243769124150276,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.0492146015167236,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.02997511625289917,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.0772786140441895,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.02930292673408985,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.055750608444214,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.03037521243095398,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.033710479736328,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.029957240447402,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.0394721031188965,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.02925589308142662,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.0878353118896484,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.03003518097102642,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.048628330230713,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.031545430421829224,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.0775623321533203,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.031310439109802246,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.01645565032959,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.03074057213962078,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.0651440620422363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.028704499825835228,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.0527148246765137,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.029061976820230484,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.0854501724243164,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.03034680150449276,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.056356906890869,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.028715113177895546,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.0874247550964355,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.030256496742367744,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.052424430847168,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.02949833683669567,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.0663650035858154,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.028980428352952003,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.0700113773345947,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.028526026755571365,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.0856711864471436,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.029572727158665657,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.0862035751342773,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.02955584041774273,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.030802011489868,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.02881692908704281,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.074392318725586,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.029564999043941498,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.057894229888916,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.030315004289150238,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.096233367919922,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.030666006729006767,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.040799617767334,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.028901133686304092,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.055051803588867,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.030359387397766113,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.085972309112549,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.03117106668651104,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.0786352157592773,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.029238386079669,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.0375771522521973,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.029324093833565712,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.1140198707580566,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.02875387854874134,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.090881109237671,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.029023943468928337,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.0755701065063477,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.0291463453322649,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.0723776817321777,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.02795177884399891,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.058993339538574,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.029499078169465065,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.0856404304504395,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.029889609664678574,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.0622942447662354,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.03107610158622265,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.114957332611084,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.02967257983982563,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.088871479034424,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.029068559408187866,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.1103081703186035,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.029334546998143196,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.0841622352600098,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.029259610921144485,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.0662848949432373,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.030072513967752457,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.0900328159332275,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.033321622759103775,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 3.0871129035949707,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.271463394165039,
+ "eval_runtime": 57.9048,
+ "eval_samples_per_second": 42.173,
+ "eval_steps_per_second": 1.33,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.03635408729314804,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 2.9576284885406494,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.03927835449576378,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 2.999624729156494,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.03371119126677513,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 2.971052646636963,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.033049341291189194,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 2.980811595916748,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.03830729424953461,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 2.9985947608947754,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.036573827266693115,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 2.9892568588256836,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.031196987256407738,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 2.9854254722595215,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.03376495838165283,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 2.9840264320373535,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.036272983998060226,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.0168871879577637,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.03328242897987366,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 2.9815053939819336,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.03158426657319069,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 2.994525909423828,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.03320910036563873,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.008638858795166,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.03160254284739494,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 2.979243040084839,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.03158991411328316,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 2.938634157180786,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.031116485595703125,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 2.9957656860351562,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.030757686123251915,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.0106635093688965,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.02992827817797661,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 2.996793746948242,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.0330309197306633,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 2.9859161376953125,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.029488038271665573,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.0081536769866943,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.03095114603638649,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.006385087966919,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.03194110840559006,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.0125250816345215,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.03306785598397255,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 2.9640283584594727,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.030883541330695152,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.0223653316497803,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.03224025294184685,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.0274910926818848,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.03182433918118477,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.047736406326294,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.03387221321463585,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 2.9821343421936035,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.03407924994826317,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.041207790374756,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.031546324491500854,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.0298609733581543,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.03354770317673683,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.0212411880493164,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.031692177057266235,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.007350444793701,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.03218817710876465,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.0019829273223877,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.03233771771192551,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.0400614738464355,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.028998777270317078,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.022221088409424,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.030422436073422432,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.0071797370910645,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.030207227915525436,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 2.9762964248657227,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.0305719543248415,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.075176239013672,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.03173545375466347,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.028982162475586,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.02799576334655285,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.0242385864257812,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.03140150383114815,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.012221336364746,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.029416466131806374,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.0037949085235596,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.028385788202285767,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.0350992679595947,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.031597770750522614,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.065139055252075,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.02926146425306797,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.014230728149414,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.02970231883227825,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.06636381149292,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.029684338718652725,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.024610996246338,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.02908361330628395,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.0093719959259033,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.029891187325119972,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.0735411643981934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.028607673943042755,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.044274091720581,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.029806701466441154,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.0449328422546387,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.03161340951919556,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.046813488006592,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.028860243037343025,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.011828660964966,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.0285879448056221,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.043640613555908,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.0296401996165514,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.071589469909668,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.02807026542723179,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.018320083618164,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.029037052765488625,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.0471763610839844,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.029226379469037056,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.037463426589966,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.029322804883122444,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.0472755432128906,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.028253905475139618,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.0265414714813232,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.028875533491373062,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.063941717147827,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.028460513800382614,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.0512683391571045,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.02951696142554283,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.0528573989868164,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.030593490228056908,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.057919979095459,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.028074532747268677,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.0263280868530273,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.02955678664147854,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.081322193145752,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.029724644497036934,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.059583902359009,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.029713226482272148,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.018678903579712,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.029758386313915253,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.0379555225372314,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.028971998021006584,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.041267156600952,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.02968725562095642,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.0675506591796875,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.029876509681344032,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.051586151123047,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.02853119559586048,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.0335168838500977,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.03374241292476654,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 3.027909278869629,
+ "step": 2520
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0743646764662784e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2520/training_args.bin b/runs/l2r90-baseline/checkpoint-2520/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2520/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja b/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-2880/config.json b/runs/l2r90-baseline/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-2880/generation_config.json b/runs/l2r90-baseline/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-2880/model.safetensors b/runs/l2r90-baseline/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..cbb2d4c9c0c87889c2a44418aa3e5c95e013a74e
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:516c4b11ac1bff4c61f8f90e9544c9b2e5a70b2b0a2026afcd2343104a0742b7
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-2880/optimizer.pt b/runs/l2r90-baseline/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..74f8275f45f0188f970a6fd417cc9709275e954d
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6e5bc533ab14e34fccd687fe60d700ddf357ff476ef17cd6525caf7beaa1b701
+size 1166827898
diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0f2f3f214d81f9b4c1f52472b6c90d59a40880d9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9d02dbaa2c75ede2796f9723198b9e46b03b4e1acd7596aef25899b19685b472
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..421d49be31deb98414023cc92e42baeafe9660da
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c1b031e812ec63375a07102faa7694fd9d3b4de75bec687a01240ccef400a542
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..10dca30de14d8c99876de5aa078e8c5fe6e4805c
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8d37e0894989a7ddb416a70e836d2c09c377786a724fba5c284f1b245a61a12b
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5b94d5903005d6524a67bfcf0219123f4e2e99df
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3c9e6be7a3ea88e462b65bb71838108dcecb9f6c8bdaf41a50c1130618751a09
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-2880/scheduler.pt b/runs/l2r90-baseline/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5c7f6b411db3da6cc58af5b4530e84dddd5abe00
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2e88932b62643ef7a711b903ac04665eab5638dccef14b4bec61d20b2791e67b
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-2880/tokenizer.json b/runs/l2r90-baseline/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2880/trainer_state.json b/runs/l2r90-baseline/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..9c8d559ac440a4458709b238f3047f9de6460966
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20498 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.078423500061035,
+ "eval_runtime": 57.3189,
+ "eval_samples_per_second": 42.604,
+ "eval_steps_per_second": 1.343,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.02624760940670967,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.261288642883301,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.03552987053990364,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.2682106494903564,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.040380869060754776,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.2813782691955566,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.03861293941736221,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.258193016052246,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.04091731086373329,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.3052220344543457,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.043269723653793335,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.2820217609405518,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.04580320790410042,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.312005043029785,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.047336459159851074,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.3217713832855225,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.04587360844016075,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.2869949340820312,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.04002532362937927,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.2879223823547363,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.0368693508207798,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.2983577251434326,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.037317775189876556,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.3062524795532227,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.03655700013041496,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.31072735786438,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.033466216176748276,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.2954514026641846,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.030785970389842987,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.27952241897583,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.029669342562556267,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.2613229751586914,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.033505845814943314,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.2987258434295654,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.032395437359809875,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.296708583831787,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.031322430819272995,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.3160252571105957,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.029345186427235603,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.307143211364746,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.026715675368905067,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.286012649536133,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.028592750430107117,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.2528645992279053,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.028861764818429947,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.2795748710632324,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.02609528973698616,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.326317071914673,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.027718454599380493,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.268110513687134,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.026382867246866226,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.294499397277832,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.027083583176136017,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.3202130794525146,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.02516467310488224,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.3485021591186523,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.024148069322109222,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.3099541664123535,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.024552857503294945,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.2885942459106445,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.02551114559173584,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.3232040405273438,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.025423647835850716,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.3097832202911377,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.023211417719721794,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.3075966835021973,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.024623898789286613,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.2917377948760986,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.02677663415670395,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.356265068054199,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.02673642337322235,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.314903736114502,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.024466225877404213,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.30912446975708,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.025370748713612556,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.363640069961548,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.026536114513874054,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.304851531982422,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.026536893099546432,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.342902898788452,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.02645307406783104,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.3164432048797607,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.02523060515522957,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.34376859664917,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.02590431459248066,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.3479061126708984,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.0230655986815691,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.2890982627868652,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.02403893694281578,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.345456600189209,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.02238907851278782,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.3471767902374268,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.021377569064497948,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.319232225418091,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.02272353693842888,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.316892623901367,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.02413824200630188,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.3450968265533447,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.02334682270884514,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.3640642166137695,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.021966520696878433,
+ "learning_rate": 0.000303,
+ "loss": 3.358555316925049,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.023984894156455994,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.358398914337158,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.023520268499851227,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.3224587440490723,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.0231285709887743,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.3671562671661377,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.022818170487880707,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.361027717590332,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.02280694991350174,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.2923383712768555,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.021362345665693283,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.3346734046936035,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.02372218668460846,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.354748249053955,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.022949764505028725,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.3857789039611816,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.021954255178570747,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.343330144882202,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.022438595071434975,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.328618049621582,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.021969925612211227,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.355971574783325,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.022706981748342514,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.3529577255249023,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.024083172902464867,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.348170757293701,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.023259354755282402,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.3474369049072266,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.02218157798051834,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.369053840637207,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.02305331453680992,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.3622641563415527,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.022796178236603737,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.351510524749756,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.021299205720424652,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.339406728744507,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.022786924615502357,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.3182196617126465,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.021729128435254097,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.383904218673706,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.024604283273220062,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 3.2912864685058594,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.0894880294799805,
+ "eval_runtime": 58.3967,
+ "eval_samples_per_second": 41.817,
+ "eval_steps_per_second": 1.319,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.02673362009227276,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.270986795425415,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.04036599025130272,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.2358999252319336,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.047250088304281235,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.2706189155578613,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.04642819985747337,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.236340284347534,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.039405327290296555,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.203136682510376,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.038748081773519516,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.2636070251464844,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.039233580231666565,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.2751896381378174,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.03429320082068443,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.2765679359436035,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.03267038241028786,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.2387681007385254,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.03349406272172928,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.282162666320801,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.032002877444028854,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.246623992919922,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.031509071588516235,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.2374932765960693,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.02958357334136963,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.24362850189209,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.03248635679483414,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.237210273742676,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.03230655938386917,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.251129150390625,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.029688173905014992,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.2141575813293457,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.027773940935730934,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.264249086380005,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.02687031403183937,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.2513632774353027,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.02503623627126217,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.217961072921753,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.027071228250861168,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.2784645557403564,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.02600773423910141,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.2641382217407227,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.026998162269592285,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.2646782398223877,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.025579402223229408,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.2700581550598145,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.02645333670079708,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.2545266151428223,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.02717350795865059,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.2795705795288086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.02675914391875267,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.2936904430389404,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.02751768007874489,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.2329440116882324,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.025446882471442223,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.274294376373291,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.02365768514573574,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.2585229873657227,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.0266884732991457,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.2872962951660156,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.02572469227015972,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.253828287124634,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.02540273219347,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.303926944732666,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.024694625288248062,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.264145851135254,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.027233170345425606,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.286924362182617,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02635285072028637,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.2930331230163574,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.025157805532217026,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.2906906604766846,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.024729596450924873,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.277128219604492,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.023809000849723816,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.2957217693328857,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02474421076476574,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.27791690826416,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.02439294196665287,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.2912981510162354,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.023601334542036057,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.317446231842041,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.0253264419734478,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.2742302417755127,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.02389134280383587,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.281032085418701,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.02451813407242298,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.3339295387268066,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.024324845522642136,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.290402889251709,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.024939008057117462,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.2933621406555176,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.024450505152344704,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.3055014610290527,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.02262270078063011,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.2964205741882324,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.02311946265399456,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.3102402687072754,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.024362705647945404,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.3020925521850586,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.024785032495856285,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.287290096282959,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.02358846180140972,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.334261894226074,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.024330584332346916,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.2799081802368164,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.023546984419226646,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.3228847980499268,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.023060832172632217,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.325793981552124,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.02388455718755722,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.2892062664031982,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.023740701377391815,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.2969369888305664,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.02458217367529869,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.315610408782959,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.023173579946160316,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.2561936378479004,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.024137936532497406,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.2839202880859375,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.022436272352933884,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.3057351112365723,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.023997878655791283,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.3583154678344727,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.024119896814227104,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.322892427444458,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.022283365949988365,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.3145735263824463,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.023433363065123558,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.3564529418945312,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.02224746160209179,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.2985916137695312,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.023786570876836777,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.302539587020874,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.025200961157679558,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.309976577758789,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.023849809542298317,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.343660354614258,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.023311031982302666,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.2978909015655518,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.023174891248345375,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.326188802719116,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.02681034244596958,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 3.3422017097473145,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.109009265899658,
+ "eval_runtime": 58.0506,
+ "eval_samples_per_second": 42.067,
+ "eval_steps_per_second": 1.326,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.028057295829057693,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.1825037002563477,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.034786805510520935,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.200389862060547,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.03751441463828087,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.2047512531280518,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.035590436309576035,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.1735124588012695,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.03426426649093628,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.199483871459961,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.037030890583992004,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.213322639465332,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.032468877732753754,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.179354190826416,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.027202758938074112,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.216036319732666,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.031184563413262367,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.208205223083496,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.029414812102913857,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.180370569229126,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.02679901383817196,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.210325002670288,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.02863631211221218,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.1601076126098633,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.029135366901755333,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.2262887954711914,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.030535712838172913,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.23160719871521,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.030563877895474434,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.255211353302002,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.02906505949795246,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.2006797790527344,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.030871668830513954,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.2442867755889893,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.02913421392440796,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.218442440032959,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.030087953433394432,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.2164535522460938,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.034147944301366806,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.218660831451416,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.03593966364860535,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.2240614891052246,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.032090652734041214,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.2309441566467285,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.02776424027979374,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.2702016830444336,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.027174271643161774,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.19810152053833,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.028479408472776413,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.2178421020507812,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.026820644736289978,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.2634482383728027,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.026427917182445526,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.2452914714813232,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.02679610252380371,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.243825912475586,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.026115626096725464,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.218787670135498,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.026686571538448334,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.2785303592681885,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.02685859426856041,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.2572226524353027,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.026968171820044518,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.2625553607940674,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.025397246703505516,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.262303352355957,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.025660471990704536,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.269401788711548,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.02713891491293907,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.2416529655456543,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.027591092512011528,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.2814316749572754,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.024462848901748657,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.249725103378296,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.024868478998541832,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.231574535369873,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.027015401050448418,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.282627582550049,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.02678140625357628,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.2208900451660156,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.026830527931451797,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.260575294494629,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.02430890128016472,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.2522997856140137,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.027047613635659218,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.2721078395843506,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.026861369609832764,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.2235946655273438,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.026650549843907356,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.302743434906006,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.02551819011569023,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.265763282775879,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.024843450635671616,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.272397518157959,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.024423206225037575,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.253495454788208,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.024180661886930466,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.260438919067383,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.023554982617497444,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.2342851161956787,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.023470086976885796,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.251622438430786,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.024405231699347496,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.2588815689086914,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.024147814139723778,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.2723565101623535,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.023217910900712013,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.2555792331695557,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.024921158328652382,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.26884126663208,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.02455919235944748,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.2674927711486816,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.024725932627916336,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.2511110305786133,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.025769414380192757,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.290651321411133,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.025928806513547897,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.302396297454834,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.02499072253704071,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.2813124656677246,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.025431277230381966,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.234684705734253,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.026871860027313232,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.2800331115722656,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.02578776702284813,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.3117146492004395,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.026318378746509552,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.257615089416504,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.025542527437210083,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.2969868183135986,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.025723429396748543,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.290259838104248,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.024588774889707565,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.284090280532837,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.024619122967123985,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.284879207611084,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.024441618472337723,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.283104658126831,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.02402760274708271,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.3013617992401123,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.025028521195054054,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.2892568111419678,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.029445001855492592,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 3.2474446296691895,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.128176689147949,
+ "eval_runtime": 58.8588,
+ "eval_samples_per_second": 41.489,
+ "eval_steps_per_second": 1.308,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.0298826452344656,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.138103485107422,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.0387548953294754,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.161492347717285,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.03816070780158043,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.1739673614501953,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.03409738838672638,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.182478189468384,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.033135563135147095,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.1720199584960938,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.03203624114394188,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.1699862480163574,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.031237754970788956,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.179363250732422,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.03342146426439285,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.1822280883789062,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.03561089560389519,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.1767094135284424,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.034710463136434555,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.218817949295044,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.030870331451296806,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.168602705001831,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.028004391118884087,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.224155902862549,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.030953701585531235,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.1789557933807373,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.030289432033896446,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.1890387535095215,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.030684132128953934,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.195427417755127,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.03022196516394615,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.1984965801239014,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.026935536414384842,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.2014272212982178,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.02702922560274601,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.205538034439087,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.026682600378990173,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.1954591274261475,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.0273384228348732,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.1782338619232178,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.028587577864527702,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.199565887451172,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.028848132118582726,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.181786298751831,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.027535341680049896,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.2013628482818604,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.02756025455892086,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.1749587059020996,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.027358995750546455,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.161881923675537,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.025918981060385704,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.225224256515503,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.026479976251721382,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.2129316329956055,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.02685079537332058,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.1651554107666016,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.027378706261515617,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.193108558654785,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.027240974828600883,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.190605640411377,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.028065402060747147,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.2068259716033936,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.028329014778137207,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.2316060066223145,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.028920065611600876,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.1933932304382324,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.028299354016780853,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.192711353302002,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.025996020063757896,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.243171453475952,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.0264595914632082,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.199758529663086,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.02704528719186783,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.2390713691711426,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.026571933180093765,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.219440460205078,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.025562983006238937,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.2538228034973145,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.024683184921741486,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.2372591495513916,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.025656528770923615,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.250558853149414,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.024807540699839592,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.1856563091278076,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.026149194687604904,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.172186851501465,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.02617705799639225,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.250445604324341,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.026542512699961662,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.220273494720459,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.024963509291410446,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.2330946922302246,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.027144264429807663,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.1860971450805664,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.02665683813393116,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.257209062576294,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.024872658774256706,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.2266182899475098,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.025235185399651527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.2134504318237305,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.025029495358467102,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.2150087356567383,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.025630896911025047,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.282879114151001,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.02722341939806938,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.2366206645965576,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.02522372268140316,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.2840523719787598,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.02494790591299534,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.2405319213867188,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.026662737131118774,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.239692211151123,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.026082608848810196,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.2156195640563965,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.02423974685370922,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.1863760948181152,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.025226688012480736,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.20874285697937,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.02428225241601467,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.2209601402282715,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.023819707334041595,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.2564258575439453,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.025661710649728775,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.244133949279785,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.02423049882054329,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.2376630306243896,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.02382473088800907,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.2951290607452393,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.02488875202834606,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.2848310470581055,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.02575223706662655,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.2574503421783447,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.02462848648428917,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.2464442253112793,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.024912388995289803,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.2459192276000977,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.025020884349942207,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.213622808456421,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.026330886408686638,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.2305431365966797,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.02708602324128151,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.2487688064575195,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.030908891931176186,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 3.2110159397125244,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.146673202514648,
+ "eval_runtime": 58.7379,
+ "eval_samples_per_second": 41.574,
+ "eval_steps_per_second": 1.311,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.02919941022992134,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.1533267498016357,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.03379782289266586,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.1436538696289062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.033698830753564835,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.15297532081604,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.03077736124396324,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.128848075866699,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.031112324446439743,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.1461119651794434,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.03346852958202362,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.113943099975586,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.03307225927710533,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.147498607635498,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.030230754986405373,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.1162924766540527,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.02982981689274311,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.144481658935547,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.031735632568597794,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.133862018585205,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.0312766470015049,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.110287666320801,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.029902562499046326,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.1384904384613037,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.0306564774364233,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.1138200759887695,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.03084711916744709,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.155240535736084,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.03183220326900482,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.1593515872955322,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.031826410442590714,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.1517653465270996,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.030141707509756088,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.1291589736938477,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.03158571943640709,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.178101062774658,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.03510458394885063,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.138767719268799,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.03660859912633896,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.1708762645721436,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.034004807472229004,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.1404073238372803,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.03077496401965618,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.159081220626831,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.030384650453925133,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.1872847080230713,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.02957277186214924,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.165973424911499,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.029529055580496788,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.123840808868408,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.027776138857007027,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.1705269813537598,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.028943249955773354,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.1181654930114746,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.028178401291370392,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.201678991317749,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.02865665778517723,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.18479061126709,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.029173649847507477,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.1844916343688965,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.0280170775949955,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.187105417251587,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.029103878885507584,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.183281660079956,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.029045486822724342,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.1848196983337402,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.028364380821585655,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.197348117828369,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.029651399701833725,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.1429920196533203,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.0281438659876585,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.1916513442993164,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.029437677934765816,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.172391176223755,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.03433044254779816,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.2296864986419678,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.03259492665529251,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.1962363719940186,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.0299630519002676,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.169673442840576,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.028495769947767258,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.1855409145355225,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.0290420800447464,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.23667573928833,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.03340664133429527,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.1873788833618164,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.03253626823425293,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.204796552658081,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.028142794966697693,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.193922281265259,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.028805622830986977,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.1863901615142822,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.030033612623810768,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.198482036590576,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.028154220432043076,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.215069055557251,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.028098132461309433,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.2197628021240234,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.028679659590125084,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.2062196731567383,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.029104258865118027,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.2182908058166504,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.029629778116941452,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.2165639400482178,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.028304366394877434,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.1783785820007324,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.026900488883256912,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.1876602172851562,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.027047287672758102,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.1879465579986572,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.027424108237028122,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.2407851219177246,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.029319508001208305,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.1980130672454834,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.027114372700452805,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.211357593536377,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.02579840086400509,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.2067861557006836,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.02767331339418888,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.198587417602539,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.026645265519618988,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.177441120147705,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.026181906461715698,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.226815700531006,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.027218692004680634,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.1731510162353516,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.027551919221878052,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.2268621921539307,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.025782009586691856,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.2052621841430664,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.024231620132923126,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.201770544052124,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.02474607340991497,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.1935834884643555,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.026299327611923218,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.223275899887085,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.02439914271235466,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.2513954639434814,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.02650653012096882,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.1914851665496826,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.024847298860549927,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.2173104286193848,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.028475770726799965,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 3.2184863090515137,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.168707847595215,
+ "eval_runtime": 57.4035,
+ "eval_samples_per_second": 42.541,
+ "eval_steps_per_second": 1.341,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.02837727963924408,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.119239330291748,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.03562312573194504,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.092813014984131,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.0344194732606411,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.0802855491638184,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.03188958019018173,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.1177961826324463,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.032181013375520706,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.1343281269073486,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.031085725873708725,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.118828296661377,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.031995322555303574,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.110109329223633,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.02949182130396366,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.0972743034362793,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.029299668967723846,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.1257238388061523,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.030558617785573006,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.1216022968292236,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.03181575983762741,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.093721866607666,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.030934303998947144,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.0910396575927734,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.030235569924116135,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.133371114730835,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.03350163996219635,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.1178882122039795,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.03468719869852066,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.1222469806671143,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.034203238785266876,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.12453031539917,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.03321516141295433,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.115428924560547,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.03298434242606163,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.127288341522217,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.029987169429659843,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.1194515228271484,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.031271349638700485,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.1284165382385254,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.032201699912548065,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.1912736892700195,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.0299740731716156,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.1233785152435303,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.029613154008984566,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.1535067558288574,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.03303765505552292,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.1550910472869873,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.032644059509038925,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.1367573738098145,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.031695179641246796,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.1350858211517334,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.030685821548104286,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.121593952178955,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.03091946430504322,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.113060235977173,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.0282905176281929,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.184551239013672,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.030020160600543022,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.1673004627227783,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.031114062294363976,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.162588596343994,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.03007666766643524,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.132836103439331,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.029013793915510178,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.1552720069885254,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.02784561738371849,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.189272165298462,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.02759280428290367,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.141814708709717,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.02699868381023407,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.1127257347106934,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.02782190404832363,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.193972587585449,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.027298318222165108,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.1393446922302246,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.027321161702275276,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.0926666259765625,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.026399722322821617,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.1578259468078613,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.02770969271659851,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.1816914081573486,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.028344258666038513,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.1860158443450928,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.028091782703995705,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.1585092544555664,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.026564039289951324,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.131692409515381,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.026349127292633057,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.160431385040283,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.030081138014793396,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.1647543907165527,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.029250919818878174,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.1522226333618164,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.0282477755099535,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.139249324798584,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.0284534003585577,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.154621124267578,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.02677394263446331,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.1375081539154053,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.026511570438742638,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.1355676651000977,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.02766270563006401,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.143874168395996,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.027013977989554405,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.1602487564086914,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.026629449799656868,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.1549456119537354,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.0272684209048748,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.1739020347595215,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.027645915746688843,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.172959327697754,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.025413552299141884,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.1649041175842285,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.028406929224729538,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.1821887493133545,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.027753977105021477,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.163238048553467,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.02693209983408451,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.1921019554138184,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.02671878971159458,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.159438133239746,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.027099428698420525,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.1806793212890625,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.025098824873566628,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.158310651779175,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.02655479498207569,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.1927309036254883,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.025836164131760597,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.1578540802001953,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.02657160349190235,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.1658592224121094,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.027223270386457443,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.1760241985321045,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.025695348158478737,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.202389717102051,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.025819813832640648,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.2111103534698486,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.02621106244623661,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.1560299396514893,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.026585502550005913,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.176602840423584,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.03038652054965496,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 3.1799943447113037,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.193845748901367,
+ "eval_runtime": 57.9911,
+ "eval_samples_per_second": 42.11,
+ "eval_steps_per_second": 1.328,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.03135736659169197,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.070986270904541,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.038986921310424805,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.0708835124969482,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.035115309059619904,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.1070001125335693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.032868221402168274,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.082258701324463,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.03703184798359871,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.057744264602661,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.03831101953983307,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.0789475440979004,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.036301176995038986,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.0565457344055176,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.03166409209370613,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.0574910640716553,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.02984458953142166,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.080018997192383,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.034001126885414124,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.0813302993774414,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.03814122825860977,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.0579185485839844,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.03870106488466263,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.0857529640197754,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.04005642607808113,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.0880727767944336,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.03760755434632301,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.0795767307281494,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.03260907158255577,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.1006722450256348,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.029866183176636696,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.0850343704223633,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.030988436192274094,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.0743892192840576,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.03698134422302246,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.056002378463745,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.03696191683411598,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.069124698638916,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.03453022986650467,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.1017374992370605,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.032458193600177765,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.1166419982910156,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.02892368845641613,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.0863287448883057,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.030625877901911736,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.0938801765441895,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.030280394479632378,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.086432933807373,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.029181072488427162,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.0977859497070312,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.028039127588272095,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.112367868423462,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.02910272404551506,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.1253533363342285,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.03086872771382332,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.103689670562744,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.02883610874414444,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.111599922180176,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.028490116819739342,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.1410303115844727,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.028714407235383987,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.133023738861084,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.029553238302469254,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.107003688812256,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.02997349575161934,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.135680675506592,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.029087591916322708,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.1103882789611816,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.029947761446237564,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.0961623191833496,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.02950555644929409,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.1162495613098145,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.029569562524557114,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.096372127532959,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.031366728246212006,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.120089530944824,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.0313473716378212,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.100785732269287,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.028021851554512978,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.0940558910369873,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.027477117255330086,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.1210484504699707,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.029690749943256378,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.1350059509277344,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.029659733176231384,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.1220788955688477,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.028260033577680588,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.1000189781188965,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.027554461732506752,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.1044349670410156,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.027929967269301414,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.1353745460510254,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.029309319332242012,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.1578292846679688,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.029183287173509598,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.1217150688171387,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.02707512304186821,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.119617223739624,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.03006536327302456,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.1481480598449707,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.03026784397661686,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.142770767211914,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.02827967144548893,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.1582541465759277,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.028892774134874344,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.1530709266662598,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.03023783676326275,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.1102521419525146,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.027951840311288834,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.139113426208496,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.02807226963341236,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.149428367614746,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.029721451923251152,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.120518207550049,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.02867680788040161,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.1109931468963623,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.02663845382630825,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.116053581237793,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.028366081416606903,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.1444268226623535,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.028931576758623123,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.163870096206665,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.02664567157626152,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.1349146366119385,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.028031354770064354,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.1638219356536865,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.029421644285321236,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.1601340770721436,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.027983693405985832,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.1636788845062256,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.02719871513545513,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.1246869564056396,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.03030397929251194,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.1749579906463623,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.029340893030166626,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.1751790046691895,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.027624106034636497,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.178704023361206,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.02830546908080578,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.177117347717285,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.03168068826198578,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.125427722930908,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.033327918499708176,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 3.13071870803833,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.218388080596924,
+ "eval_runtime": 57.611,
+ "eval_samples_per_second": 42.388,
+ "eval_steps_per_second": 1.337,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.03336749225854874,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.064338445663452,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.038479577749967575,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.0478687286376953,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.032562077045440674,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.019803762435913,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.0331728458404541,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.0599141120910645,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.03361169621348381,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.0315823554992676,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.032973092049360275,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.025928258895874,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.030388766899704933,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.070671558380127,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.030899515375494957,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.012938976287842,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.0335412323474884,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.0156331062316895,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.035731345415115356,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.0578935146331787,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.03379053995013237,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.048701286315918,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.031242679804563522,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.0832977294921875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.030997611582279205,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.0663583278656006,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.03397331386804581,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.0115303993225098,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.03349050134420395,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.02864933013916,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.031079065054655075,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.053401470184326,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.029346197843551636,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.0700926780700684,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.030784208327531815,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.066525936126709,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.03368140384554863,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.0956242084503174,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.03356792405247688,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.0394320487976074,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.029909325763583183,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.031525135040283,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.030711578205227852,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.075474977493286,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.03452072665095329,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.0228147506713867,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.03235075995326042,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.0664501190185547,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.029671598225831985,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.055302143096924,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.033868469297885895,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.053016185760498,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.0329158715903759,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.1017332077026367,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.031983695924282074,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.0706472396850586,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.02963697351515293,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.0711426734924316,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.03162391856312752,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.044867515563965,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.030773654580116272,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.0963196754455566,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.031220417469739914,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.0892245769500732,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.030928561463952065,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.1181323528289795,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.03192073106765747,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.066584348678589,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.03075495734810829,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.0436511039733887,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.02895325981080532,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.123373031616211,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.029738714918494225,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.1049060821533203,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.031937118619680405,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.099111318588257,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.029886579141020775,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.0970301628112793,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.029569782316684723,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.062499761581421,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.02924283780157566,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.0618538856506348,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.030955800786614418,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.0745606422424316,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.03145027533173561,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.122938632965088,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.028926467522978783,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.0808773040771484,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.029592076316475868,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.1167049407958984,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.028070243075489998,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.0781636238098145,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.028642063960433006,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.093937397003174,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.0290757454931736,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.1079559326171875,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.027867283672094345,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.0885753631591797,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.02896495908498764,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.0889246463775635,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.02777806483209133,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.084439277648926,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.02904277853667736,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.125246047973633,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.029010627418756485,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.111422061920166,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.028178894892334938,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.115067481994629,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.02763799950480461,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.1228506565093994,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.028052521869540215,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.1150546073913574,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.029438253492116928,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.0921788215637207,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.028195543214678764,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.1408450603485107,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.02937602624297142,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.111612319946289,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.028430236503481865,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.093441963195801,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.027624569833278656,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.1184372901916504,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.029797039926052094,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.1074559688568115,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.028728218749165535,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.161461114883423,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.02704412117600441,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.13325834274292,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.028067929670214653,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.093172550201416,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.028848597779870033,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.1461825370788574,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.02956785447895527,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.141045570373535,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.028214773163199425,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.136033773422241,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.029155246913433075,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.0986382961273193,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.02968468703329563,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.11630916595459,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.028356682509183884,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.1134822368621826,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.03301718086004257,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 3.118591785430908,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.246211528778076,
+ "eval_runtime": 57.5155,
+ "eval_samples_per_second": 42.458,
+ "eval_steps_per_second": 1.339,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.033637888729572296,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.0192506313323975,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.04057888314127922,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 2.977440595626831,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.03579854592680931,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 2.9902219772338867,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.03131850063800812,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 2.9988656044006348,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.03262636810541153,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.037747621536255,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.03260403499007225,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 2.987508773803711,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.033165134489536285,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.0511183738708496,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.03303489834070206,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.0083041191101074,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.03185006603598595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 2.9966228008270264,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.03135651350021362,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.036759376525879,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.03216925635933876,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.0304150581359863,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.03231671452522278,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.0300474166870117,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.03263348713517189,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.0298163890838623,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.032594650983810425,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 2.98789119720459,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.03406650573015213,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.0102760791778564,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.03379780799150467,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.0564732551574707,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.03172587603330612,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.0410943031311035,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.02938298135995865,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.0517544746398926,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.03206252679228783,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.0352776050567627,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.033386752009391785,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.020535945892334,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.033984702080488205,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.0552730560302734,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0329931266605854,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.0075573921203613,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.032101184129714966,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.046388864517212,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.032873060554265976,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.0340287685394287,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.030672110617160797,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.0347838401794434,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.030014054849743843,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.0315332412719727,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.02932834066450596,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.084340810775757,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.029181934893131256,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.032886028289795,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.02966221794486046,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.032766819000244,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.029713116586208344,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.028778076171875,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.029243769124150276,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.0492146015167236,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.02997511625289917,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.0772786140441895,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.02930292673408985,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.055750608444214,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.03037521243095398,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.033710479736328,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.029957240447402,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.0394721031188965,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.02925589308142662,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.0878353118896484,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.03003518097102642,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.048628330230713,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.031545430421829224,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.0775623321533203,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.031310439109802246,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.01645565032959,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.03074057213962078,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.0651440620422363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.028704499825835228,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.0527148246765137,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.029061976820230484,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.0854501724243164,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.03034680150449276,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.056356906890869,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.028715113177895546,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.0874247550964355,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.030256496742367744,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.052424430847168,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.02949833683669567,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.0663650035858154,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.028980428352952003,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.0700113773345947,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.028526026755571365,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.0856711864471436,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.029572727158665657,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.0862035751342773,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.02955584041774273,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.030802011489868,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.02881692908704281,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.074392318725586,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.029564999043941498,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.057894229888916,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.030315004289150238,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.096233367919922,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.030666006729006767,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.040799617767334,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.028901133686304092,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.055051803588867,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.030359387397766113,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.085972309112549,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.03117106668651104,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.0786352157592773,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.029238386079669,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.0375771522521973,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.029324093833565712,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.1140198707580566,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.02875387854874134,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.090881109237671,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.029023943468928337,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.0755701065063477,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.0291463453322649,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.0723776817321777,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.02795177884399891,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.058993339538574,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.029499078169465065,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.0856404304504395,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.029889609664678574,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.0622942447662354,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.03107610158622265,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.114957332611084,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.02967257983982563,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.088871479034424,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.029068559408187866,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.1103081703186035,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.029334546998143196,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.0841622352600098,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.029259610921144485,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.0662848949432373,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.030072513967752457,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.0900328159332275,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.033321622759103775,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 3.0871129035949707,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.271463394165039,
+ "eval_runtime": 57.9048,
+ "eval_samples_per_second": 42.173,
+ "eval_steps_per_second": 1.33,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.03635408729314804,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 2.9576284885406494,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.03927835449576378,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 2.999624729156494,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.03371119126677513,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 2.971052646636963,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.033049341291189194,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 2.980811595916748,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.03830729424953461,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 2.9985947608947754,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.036573827266693115,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 2.9892568588256836,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.031196987256407738,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 2.9854254722595215,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.03376495838165283,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 2.9840264320373535,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.036272983998060226,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.0168871879577637,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.03328242897987366,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 2.9815053939819336,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.03158426657319069,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 2.994525909423828,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.03320910036563873,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.008638858795166,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.03160254284739494,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 2.979243040084839,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.03158991411328316,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 2.938634157180786,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.031116485595703125,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 2.9957656860351562,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.030757686123251915,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.0106635093688965,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.02992827817797661,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 2.996793746948242,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.0330309197306633,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 2.9859161376953125,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.029488038271665573,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.0081536769866943,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.03095114603638649,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.006385087966919,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.03194110840559006,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.0125250816345215,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.03306785598397255,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 2.9640283584594727,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.030883541330695152,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.0223653316497803,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.03224025294184685,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.0274910926818848,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.03182433918118477,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.047736406326294,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.03387221321463585,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 2.9821343421936035,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.03407924994826317,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.041207790374756,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.031546324491500854,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.0298609733581543,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.03354770317673683,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.0212411880493164,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.031692177057266235,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.007350444793701,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.03218817710876465,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.0019829273223877,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.03233771771192551,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.0400614738464355,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.028998777270317078,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.022221088409424,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.030422436073422432,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.0071797370910645,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.030207227915525436,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 2.9762964248657227,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.0305719543248415,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.075176239013672,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.03173545375466347,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.028982162475586,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.02799576334655285,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.0242385864257812,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.03140150383114815,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.012221336364746,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.029416466131806374,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.0037949085235596,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.028385788202285767,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.0350992679595947,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.031597770750522614,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.065139055252075,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.02926146425306797,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.014230728149414,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.02970231883227825,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.06636381149292,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.029684338718652725,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.024610996246338,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.02908361330628395,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.0093719959259033,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.029891187325119972,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.0735411643981934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.028607673943042755,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.044274091720581,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.029806701466441154,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.0449328422546387,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.03161340951919556,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.046813488006592,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.028860243037343025,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.011828660964966,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.0285879448056221,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.043640613555908,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.0296401996165514,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.071589469909668,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.02807026542723179,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.018320083618164,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.029037052765488625,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.0471763610839844,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.029226379469037056,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.037463426589966,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.029322804883122444,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.0472755432128906,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.028253905475139618,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.0265414714813232,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.028875533491373062,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.063941717147827,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.028460513800382614,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.0512683391571045,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.02951696142554283,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.0528573989868164,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.030593490228056908,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.057919979095459,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.028074532747268677,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.0263280868530273,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.02955678664147854,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.081322193145752,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.029724644497036934,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.059583902359009,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.029713226482272148,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.018678903579712,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.029758386313915253,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.0379555225372314,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.028971998021006584,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.041267156600952,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.02968725562095642,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.0675506591796875,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.029876509681344032,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.051586151123047,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.02853119559586048,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.0335168838500977,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.03374241292476654,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 3.027909278869629,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.030695006251335144,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 2.9685583114624023,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0339728482067585,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 2.936551570892334,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.03372679278254509,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 2.9631872177124023,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.03172554448246956,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 2.969510078430176,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.030516354367136955,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 2.930342674255371,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.03261839970946312,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 2.9920477867126465,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.030618635937571526,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 2.9693596363067627,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.030576011165976524,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 2.929804801940918,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.03224785998463631,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 2.95280122756958,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.030605586245656013,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 2.931358575820923,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.031206639483571053,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 2.962998151779175,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.031317099928855896,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 2.9706358909606934,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.031614501029253006,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 2.9326977729797363,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.031002728268504143,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 2.99625301361084,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.03216253221035004,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 2.920912027359009,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.033380527049303055,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 2.9688234329223633,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.03085758350789547,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 2.9811222553253174,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.03258587792515755,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 2.932366371154785,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.031639691442251205,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 2.988865613937378,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.03144839406013489,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 2.9750723838806152,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.030864672735333443,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 2.999329090118408,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.030679022893309593,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 2.974264144897461,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.03236627206206322,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 2.9808404445648193,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.03115275874733925,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.0216290950775146,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.029868317767977715,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 2.9739654064178467,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.03446527197957039,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 2.9565515518188477,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.033536117523908615,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 2.958047389984131,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.03113192319869995,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 2.9662246704101562,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.03089931420981884,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 2.9972119331359863,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.030822310596704483,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.020500421524048,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.032264843583106995,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 2.9895882606506348,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.031226349994540215,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 2.958217144012451,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.03110436722636223,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 2.9884791374206543,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.03358910232782364,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 2.9908103942871094,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.03348501771688461,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 2.985677480697632,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.029595404863357544,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 2.979318380355835,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.031459368765354156,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 2.9921798706054688,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.03067564032971859,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 2.977778434753418,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.030506785959005356,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 2.9841768741607666,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.029978664591908455,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 2.995443344116211,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.030900927260518074,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.0177016258239746,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.029710259288549423,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.01532244682312,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.031190430745482445,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 2.9791834354400635,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.03150483965873718,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 2.998800754547119,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.030249621719121933,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.0187838077545166,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.031052686274051666,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 2.984062433242798,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.02961024083197117,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 2.970818042755127,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.03020382858812809,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 2.9913196563720703,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.031914591789245605,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 2.993145704269409,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.03061537817120552,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.0131571292877197,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.029960963875055313,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.0017619132995605,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.030348235741257668,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.0220556259155273,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.03030729480087757,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.0499038696289062,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.030809229239821434,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 2.986651659011841,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.029438305646181107,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.0351314544677734,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.029887216165661812,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.021660327911377,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.02955588512122631,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 2.9990694522857666,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.029586469754576683,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.0296952724456787,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.030632557347416878,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.0124430656433105,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.030579863116145134,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.0381762981414795,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.03282996267080307,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.0329041481018066,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.031176412478089333,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.037598133087158,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.028640294447541237,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.0004701614379883,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.0326533280313015,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.0328853130340576,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.03041139990091324,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.0149893760681152,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.030366195365786552,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.0379695892333984,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.03106893040239811,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.0284576416015625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.0288887657225132,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.028569221496582,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.03154266998171806,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.0737850666046143,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.0305064395070076,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.0744194984436035,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.030967431142926216,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.0170607566833496,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.036945272237062454,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 3.0355610847473145,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.329805374145508,
+ "eval_runtime": 64.1339,
+ "eval_samples_per_second": 38.077,
+ "eval_steps_per_second": 1.201,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.03236053138971329,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 2.9230728149414062,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.0395631417632103,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 2.92578125,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.0361955352127552,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 2.937826633453369,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.035918429493904114,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 2.9481167793273926,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.03309318795800209,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 2.8919506072998047,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0327315591275692,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 2.903585433959961,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.033141329884529114,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 2.942802667617798,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.033966805785894394,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 2.9210848808288574,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.03218021243810654,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 2.933588981628418,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.03272828459739685,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 2.9202046394348145,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.030764421448111534,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 2.916121006011963,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.031897611916065216,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 2.957679033279419,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.03199796751141548,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 2.9276621341705322,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.030233927071094513,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 2.9006078243255615,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.03035062551498413,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 2.920341730117798,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.03029404580593109,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 2.937875747680664,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.029371997341513634,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 2.9832730293273926,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.030163884162902832,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 2.94118332862854,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.03072655200958252,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 2.9632580280303955,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.02975962497293949,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 2.9346470832824707,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.029814405366778374,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 2.9451403617858887,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.030135225504636765,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 2.944918155670166,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.029395751655101776,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 2.925586223602295,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.0306396484375,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 2.9487359523773193,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.03110465221107006,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 2.963491678237915,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.03155078738927841,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 2.9355576038360596,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.030103784054517746,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 2.953045606613159,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.030827803537249565,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 2.968574047088623,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.0310911163687706,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 2.977902412414551,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.030260782688856125,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 2.94873309135437,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.030654428526759148,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 2.9282753467559814,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.02993479184806347,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 2.97076416015625,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.030244970694184303,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 2.960441827774048,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.030003050342202187,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 2.9672164916992188,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.030057862401008606,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 2.961918830871582,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.029379919171333313,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 2.9918394088745117,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.03060895763337612,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 2.94891619682312,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.030228130519390106,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 2.9726617336273193,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.029833678156137466,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 2.9494335651397705,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.030842263251543045,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 2.9580483436584473,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.029707176610827446,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 2.963684320449829,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.030966786667704582,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 2.9710958003997803,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.03137032687664032,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.003589391708374,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.03123190440237522,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 2.9742860794067383,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.030608588829636574,
+ "learning_rate": 0.000110418175419276,
+ "loss": 2.97404408454895,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.030018644407391548,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 2.972231864929199,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.03056846372783184,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 2.9871716499328613,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.02888435684144497,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 2.9774169921875,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.03173990175127983,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 2.941751480102539,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.030047327280044556,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 2.990516185760498,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.030729815363883972,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 2.9745054244995117,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.030600983649492264,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 2.9676642417907715,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.03208782523870468,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.006941556930542,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.031162556260824203,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 2.9740395545959473,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.031870320439338684,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 2.9809296131134033,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.03184300288558006,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 2.9781875610351562,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.029629364609718323,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 2.9912891387939453,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.03322628512978554,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 2.9957847595214844,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.0322413295507431,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 2.9894957542419434,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.030753307044506073,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.0320589542388916,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.03447900339961052,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.0189106464385986,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.032016605138778687,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.0162312984466553,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.030838409438729286,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.000263214111328,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.034556787461042404,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 2.996680736541748,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.03266173601150513,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 2.990510940551758,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.030973197892308235,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 2.9767255783081055,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.03442178666591644,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 2.9941515922546387,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.03215927258133888,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.006618022918701,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.030949831008911133,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.004136085510254,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.03256693854928017,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.0229580402374268,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.02986278384923935,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 2.9580869674682617,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.03597614914178848,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.995757579803467,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.359405517578125,
+ "eval_runtime": 68.1721,
+ "eval_samples_per_second": 35.821,
+ "eval_steps_per_second": 1.129,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.03432514891028404,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 2.922243118286133,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.0345693901181221,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 2.9114043712615967,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0349399633705616,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 2.8668057918548584,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.03161955252289772,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 2.8996763229370117,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.032124556601047516,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 2.8662075996398926,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.03291172534227371,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 2.9088220596313477,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.033552397042512894,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 2.9076967239379883,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.03218761831521988,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 2.8726117610931396,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.03289451450109482,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 2.9127345085144043,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.033019524067640305,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 2.920897960662842,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.03153533488512039,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 2.9124374389648438,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.032293178141117096,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 2.877277374267578,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.031969670206308365,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 2.921635150909424,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.03023488074541092,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 2.9336061477661133,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.03215106576681137,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 2.905877113342285,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.030415233224630356,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 2.909331798553467,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.031150560826063156,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 2.9302995204925537,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.03152180463075638,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 2.9171547889709473,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.02996763028204441,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 2.916459798812866,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.030544105917215347,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 2.898505449295044,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.030963370576500893,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 2.948704719543457,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.03191790357232094,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 2.9182960987091064,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.02956460975110531,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 2.9211082458496094,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.031227655708789825,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 2.942269802093506,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.031859587877988815,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 2.9169349670410156,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.03045276179909706,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 2.9240684509277344,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.031577471643686295,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 2.9135894775390625,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.031031791120767593,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 2.913355827331543,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.030012110248208046,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 2.8860886096954346,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.029772799462080002,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 2.974332809448242,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.029961293563246727,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 2.943694591522217,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.029537590220570564,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 2.93180513381958,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.030834559351205826,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 2.9442358016967773,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.02998523600399494,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 2.9336910247802734,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.030030962079763412,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 2.93593692779541,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.030250679701566696,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 2.9370553493499756,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.02942339889705181,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 2.9406495094299316,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.030448907986283302,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 2.9552698135375977,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.03078317642211914,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 2.933631420135498,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.029188675805926323,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 2.944002628326416,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.03153637796640396,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 2.9708166122436523,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.030299564823508263,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 2.9570717811584473,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.03169458732008934,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 2.942774772644043,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.031527288258075714,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 2.9711742401123047,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.0314360186457634,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 2.9509239196777344,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.032102152705192566,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 2.9443840980529785,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.03111223131418228,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 2.972594738006592,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.0322272814810276,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 2.9361982345581055,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.033640168607234955,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 2.9726009368896484,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.02979348786175251,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 2.9978041648864746,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.030464721843600273,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 2.949810028076172,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.03170578181743622,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 2.9677181243896484,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.02996704913675785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 2.974087953567505,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.030277637764811516,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 2.95644211769104,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.030859284102916718,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 2.957258701324463,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.029737796634435654,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 2.9502854347229004,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.029822375625371933,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 2.9625749588012695,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.030455606058239937,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 2.9669153690338135,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.029491793364286423,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 2.952991247177124,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.030326122418045998,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 2.978355646133423,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.03011198714375496,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 2.976996421813965,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.029053322970867157,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 2.9754371643066406,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.02925761416554451,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 2.94271183013916,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.029678862541913986,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 2.9386143684387207,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.031623054295778275,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 2.9660329818725586,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.030133802443742752,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 2.9580249786376953,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.029482144862413406,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 2.944474697113037,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.030708568170666695,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 2.967862367630005,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.028793595731258392,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 2.9581642150878906,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.03083081543445587,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 2.957261323928833,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.030180038884282112,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 2.9579484462738037,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.034881673753261566,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.9007070064544678,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.38953161239624,
+ "eval_runtime": 71.4964,
+ "eval_samples_per_second": 34.156,
+ "eval_steps_per_second": 1.077,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0335528701543808,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 2.8542537689208984,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.03163984417915344,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 2.864593505859375,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.034005917608737946,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 2.855072498321533,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.03235425800085068,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 2.8504371643066406,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.03171464428305626,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 2.8924291133880615,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.0333196297287941,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 2.8764119148254395,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.03050111047923565,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 2.874119281768799,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.032114267349243164,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 2.9059693813323975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.03141588345170021,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 2.879429340362549,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.03201865404844284,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 2.898519277572632,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.03152940049767494,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 2.866131544113159,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.032520636916160583,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 2.888209819793701,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.03244909271597862,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 2.888913631439209,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.03160687908530235,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 2.905120372772217,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.031518299132585526,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 2.8985815048217773,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.0304370429366827,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 2.8531291484832764,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.03239843621850014,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 2.876098394393921,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.03194732964038849,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 2.908271074295044,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.03145617991685867,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 2.874128580093384,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.031182898208498955,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 2.894904613494873,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.0323818176984787,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 2.9005861282348633,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.030969973653554916,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 2.9236338138580322,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.03238993510603905,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 2.8811020851135254,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.030876649543642998,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 2.8802058696746826,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.032251305878162384,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 2.882112503051758,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.03334304690361023,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 2.9290661811828613,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.03299225866794586,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 2.908268928527832,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.03335456922650337,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 2.9156572818756104,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.030786098912358284,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 2.8981709480285645,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.03379470854997635,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 2.9483864307403564,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.03020016849040985,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 2.908958911895752,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.032774876803159714,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 2.899505138397217,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.030589981004595757,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 2.9257779121398926,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.03366618603467941,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 2.8749284744262695,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.03038695827126503,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 2.9127252101898193,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.03400666266679764,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 2.922375202178955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.031702589243650436,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 2.8886446952819824,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.03223840892314911,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 2.906278610229492,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.0327419750392437,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 2.9436042308807373,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.030698910355567932,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 2.897615909576416,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.03250030800700188,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 2.928028106689453,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.030623450875282288,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 2.9268975257873535,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.030981825664639473,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 2.938185691833496,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.03139727562665939,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 2.9397618770599365,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.029906271025538445,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 2.921111583709717,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.03110392577946186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 2.899641513824463,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.03060143068432808,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 2.922635078430176,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.03061371110379696,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 2.925189733505249,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.03080526366829872,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 2.90329647064209,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.030412547290325165,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 2.9444844722747803,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.03024943359196186,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 2.9165825843811035,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.030653616413474083,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 2.9171371459960938,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.03027702495455742,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 2.9399702548980713,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.030466679483652115,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 2.888066291809082,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.02997267246246338,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 2.906545877456665,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.03026283159852028,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 2.9359335899353027,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.0293166171759367,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 2.9118804931640625,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.031117571517825127,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 2.968198299407959,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.03041299246251583,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 2.9333484172821045,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.03037327155470848,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 2.895859956741333,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.03161945194005966,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 2.9491944313049316,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.031058384105563164,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 2.926038980484009,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.030037537217140198,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 2.941394805908203,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.030748063698410988,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 2.998904228210449,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.030324939638376236,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 2.9226560592651367,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.030745431780815125,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 2.977520227432251,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.02952338382601738,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 2.9365828037261963,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.030132673680782318,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 2.9574036598205566,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.030379803851246834,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 2.9329113960266113,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.029670827090740204,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 2.951117515563965,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.029823293909430504,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 2.9586987495422363,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.03546741604804993,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.9378662109375,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.417913436889648,
+ "eval_runtime": 69.6157,
+ "eval_samples_per_second": 35.078,
+ "eval_steps_per_second": 1.106,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.033859945833683014,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 2.8493661880493164,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.032931916415691376,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 2.844602108001709,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.03232948109507561,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 2.8463873863220215,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.032815027981996536,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 2.846658706665039,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.03269147500395775,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 2.8286149501800537,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.03337012976408005,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 2.9202351570129395,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.031733594834804535,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 2.8601999282836914,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.032877493649721146,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 2.8566389083862305,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.033852867782115936,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 2.859797477722168,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.03143248334527016,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 2.8411474227905273,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.033198583871126175,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 2.888702392578125,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.029946891590952873,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 2.8332901000976562,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.031660862267017365,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 2.8255481719970703,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.030189910903573036,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 2.8664190769195557,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.03273668885231018,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 2.886531114578247,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.030584124848246574,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 2.872990131378174,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.032543931156396866,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 2.891528606414795,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.031016048043966293,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 2.8759872913360596,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.03190434351563454,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 2.8757526874542236,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.030996622517704964,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 2.862654447555542,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.03289523348212242,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 2.9064857959747314,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.03322640061378479,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 2.8921048641204834,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.0303682629019022,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 2.886458396911621,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.033412594348192215,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 2.8638253211975098,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.030457086861133575,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 2.835578441619873,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.03225106745958328,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 2.8652632236480713,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.03178466483950615,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 2.8796353340148926,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.03075973503291607,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 2.895150661468506,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.02984696812927723,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 2.882878303527832,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.030945835635066032,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 2.898838520050049,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.030904723331332207,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 2.885678291320801,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.030709490180015564,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 2.8645198345184326,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.030528701841831207,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 2.8877408504486084,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.030471473932266235,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 2.9136905670166016,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.030309896916151047,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 2.9238810539245605,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.030605493113398552,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 2.8624460697174072,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.029686832800507545,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 2.9085543155670166,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.030039243400096893,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 2.8624000549316406,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.02953776717185974,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 2.8893938064575195,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.029651308432221413,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 2.914975166320801,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.030902806669473648,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 2.887071132659912,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.02948773466050625,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 2.8893041610717773,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.03049607202410698,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 2.902736186981201,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.030077926814556122,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 2.8936967849731445,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.0303986556828022,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 2.9223172664642334,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.030650658532977104,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 2.9140920639038086,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.029562659561634064,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 2.918179988861084,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.03063463792204857,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 2.9222826957702637,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.030284354463219643,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 2.8685007095336914,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.029791876673698425,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 2.914102554321289,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.030152756720781326,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 2.8890669345855713,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.02957645244896412,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 2.8851962089538574,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.02963421493768692,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 2.9012060165405273,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.029279837384819984,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 2.949984073638916,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.0302849430590868,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 2.9049415588378906,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.029432371258735657,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 2.8919944763183594,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.030542058870196342,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 2.9168801307678223,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.02911970019340515,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 2.9342129230499268,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.030335791409015656,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 2.918611764907837,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.03031272068619728,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 2.8769750595092773,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.030200788751244545,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 2.8994460105895996,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.03032573312520981,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 2.920720100402832,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.03180326521396637,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 2.891145706176758,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.029451677575707436,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 2.9057083129882812,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.030261319130659103,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 2.9132461547851562,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.031257566064596176,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 2.9089136123657227,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.03043205663561821,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 2.912537097930908,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.030708536505699158,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 2.8841137886047363,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.03044111281633377,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 2.921257257461548,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.029576802626252174,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 2.908468723297119,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.03031427040696144,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 2.936445713043213,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.036312974989414215,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.8997855186462402,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2278453445328896e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-2880/training_args.bin b/runs/l2r90-baseline/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja b/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-3240/config.json b/runs/l2r90-baseline/checkpoint-3240/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-3240/generation_config.json b/runs/l2r90-baseline/checkpoint-3240/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-3240/model.safetensors b/runs/l2r90-baseline/checkpoint-3240/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..1add19f2f1040d4b57eabaa78be392881ab265c2
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a2d5eaf42372ea6e71f162be49d838261a729973ef8e95f101e7c686675e174b
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-3240/optimizer.pt b/runs/l2r90-baseline/checkpoint-3240/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..db4f028221c1446eb4685ff92d34bee00f323160
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:23c38996c8869706cb8694119a4e7734e44af55bcabd4ae2508ccd27b24f1d20
+size 1166827898
diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..03a3cf3ac5f2494e2e8da8d0950115c2ccc72e58
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:87ef313648fca4db6856222ba9d5b377ce9c78e1b0b773915470c174d8c1c2c1
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1d17deed2fc60410da20d54a2fbecb4d346413b2
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:667ef72d8ecb4d5779e98d21e507df93f664be0f5a602938961f1e875275982c
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bcc6f0655376c0e1e22412736914a5726b35023a
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:79a799e6c6d9bf9f5dc43e611788e95f712fea0df4ce951955e0ef5ae1737ed7
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..76a15cd64874a20e370ab87f745baa61cd9ca761
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:063fb8860cd08a9d7dffc329c1b857f2c84f027b83d2d31f9589ce8e5d92bba5
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3240/scheduler.pt b/runs/l2r90-baseline/checkpoint-3240/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..84479f3138d6041f59e136173e080a0ddbdf81b1
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1045dafc67457b916f466f6a7399234fef3fce01fc96958c0e146dea7784e901
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-3240/tokenizer.json b/runs/l2r90-baseline/checkpoint-3240/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-3240/trainer_state.json b/runs/l2r90-baseline/checkpoint-3240/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..17949cd890aca3314385de9746009d0a7febb771
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/trainer_state.json
@@ -0,0 +1,23058 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 45.0,
+ "eval_steps": 500,
+ "global_step": 3240,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.078423500061035,
+ "eval_runtime": 57.3189,
+ "eval_samples_per_second": 42.604,
+ "eval_steps_per_second": 1.343,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.02624760940670967,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.261288642883301,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.03552987053990364,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.2682106494903564,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.040380869060754776,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.2813782691955566,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.03861293941736221,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.258193016052246,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.04091731086373329,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.3052220344543457,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.043269723653793335,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.2820217609405518,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.04580320790410042,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.312005043029785,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.047336459159851074,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.3217713832855225,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.04587360844016075,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.2869949340820312,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.04002532362937927,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.2879223823547363,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.0368693508207798,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.2983577251434326,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.037317775189876556,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.3062524795532227,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.03655700013041496,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.31072735786438,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.033466216176748276,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.2954514026641846,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.030785970389842987,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.27952241897583,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.029669342562556267,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.2613229751586914,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.033505845814943314,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.2987258434295654,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.032395437359809875,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.296708583831787,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.031322430819272995,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.3160252571105957,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.029345186427235603,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.307143211364746,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.026715675368905067,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.286012649536133,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.028592750430107117,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.2528645992279053,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.028861764818429947,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.2795748710632324,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.02609528973698616,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.326317071914673,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.027718454599380493,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.268110513687134,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.026382867246866226,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.294499397277832,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.027083583176136017,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.3202130794525146,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.02516467310488224,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.3485021591186523,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.024148069322109222,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.3099541664123535,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.024552857503294945,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.2885942459106445,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.02551114559173584,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.3232040405273438,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.025423647835850716,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.3097832202911377,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.023211417719721794,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.3075966835021973,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.024623898789286613,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.2917377948760986,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.02677663415670395,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.356265068054199,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.02673642337322235,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.314903736114502,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.024466225877404213,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.30912446975708,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.025370748713612556,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.363640069961548,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.026536114513874054,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.304851531982422,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.026536893099546432,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.342902898788452,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.02645307406783104,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.3164432048797607,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.02523060515522957,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.34376859664917,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.02590431459248066,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.3479061126708984,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.0230655986815691,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.2890982627868652,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.02403893694281578,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.345456600189209,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.02238907851278782,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.3471767902374268,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.021377569064497948,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.319232225418091,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.02272353693842888,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.316892623901367,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.02413824200630188,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.3450968265533447,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.02334682270884514,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.3640642166137695,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.021966520696878433,
+ "learning_rate": 0.000303,
+ "loss": 3.358555316925049,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.023984894156455994,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.358398914337158,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.023520268499851227,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.3224587440490723,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.0231285709887743,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.3671562671661377,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.022818170487880707,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.361027717590332,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.02280694991350174,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.2923383712768555,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.021362345665693283,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.3346734046936035,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.02372218668460846,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.354748249053955,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.022949764505028725,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.3857789039611816,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.021954255178570747,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.343330144882202,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.022438595071434975,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.328618049621582,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.021969925612211227,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.355971574783325,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.022706981748342514,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.3529577255249023,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.024083172902464867,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.348170757293701,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.023259354755282402,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.3474369049072266,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.02218157798051834,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.369053840637207,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.02305331453680992,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.3622641563415527,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.022796178236603737,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.351510524749756,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.021299205720424652,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.339406728744507,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.022786924615502357,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.3182196617126465,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.021729128435254097,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.383904218673706,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.024604283273220062,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 3.2912864685058594,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.0894880294799805,
+ "eval_runtime": 58.3967,
+ "eval_samples_per_second": 41.817,
+ "eval_steps_per_second": 1.319,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.02673362009227276,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.270986795425415,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.04036599025130272,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.2358999252319336,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.047250088304281235,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.2706189155578613,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.04642819985747337,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.236340284347534,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.039405327290296555,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.203136682510376,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.038748081773519516,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.2636070251464844,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.039233580231666565,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.2751896381378174,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.03429320082068443,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.2765679359436035,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.03267038241028786,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.2387681007385254,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.03349406272172928,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.282162666320801,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.032002877444028854,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.246623992919922,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.031509071588516235,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.2374932765960693,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.02958357334136963,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.24362850189209,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.03248635679483414,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.237210273742676,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.03230655938386917,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.251129150390625,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.029688173905014992,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.2141575813293457,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.027773940935730934,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.264249086380005,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.02687031403183937,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.2513632774353027,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.02503623627126217,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.217961072921753,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.027071228250861168,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.2784645557403564,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.02600773423910141,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.2641382217407227,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.026998162269592285,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.2646782398223877,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.025579402223229408,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.2700581550598145,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.02645333670079708,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.2545266151428223,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.02717350795865059,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.2795705795288086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.02675914391875267,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.2936904430389404,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.02751768007874489,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.2329440116882324,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.025446882471442223,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.274294376373291,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.02365768514573574,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.2585229873657227,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.0266884732991457,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.2872962951660156,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.02572469227015972,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.253828287124634,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.02540273219347,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.303926944732666,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.024694625288248062,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.264145851135254,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.027233170345425606,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.286924362182617,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02635285072028637,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.2930331230163574,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.025157805532217026,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.2906906604766846,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.024729596450924873,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.277128219604492,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.023809000849723816,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.2957217693328857,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02474421076476574,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.27791690826416,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.02439294196665287,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.2912981510162354,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.023601334542036057,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.317446231842041,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.0253264419734478,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.2742302417755127,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.02389134280383587,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.281032085418701,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.02451813407242298,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.3339295387268066,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.024324845522642136,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.290402889251709,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.024939008057117462,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.2933621406555176,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.024450505152344704,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.3055014610290527,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.02262270078063011,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.2964205741882324,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.02311946265399456,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.3102402687072754,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.024362705647945404,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.3020925521850586,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.024785032495856285,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.287290096282959,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.02358846180140972,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.334261894226074,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.024330584332346916,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.2799081802368164,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.023546984419226646,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.3228847980499268,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.023060832172632217,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.325793981552124,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.02388455718755722,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.2892062664031982,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.023740701377391815,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.2969369888305664,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.02458217367529869,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.315610408782959,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.023173579946160316,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.2561936378479004,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.024137936532497406,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.2839202880859375,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.022436272352933884,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.3057351112365723,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.023997878655791283,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.3583154678344727,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.024119896814227104,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.322892427444458,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.022283365949988365,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.3145735263824463,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.023433363065123558,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.3564529418945312,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.02224746160209179,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.2985916137695312,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.023786570876836777,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.302539587020874,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.025200961157679558,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.309976577758789,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.023849809542298317,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.343660354614258,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.023311031982302666,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.2978909015655518,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.023174891248345375,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.326188802719116,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.02681034244596958,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 3.3422017097473145,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.109009265899658,
+ "eval_runtime": 58.0506,
+ "eval_samples_per_second": 42.067,
+ "eval_steps_per_second": 1.326,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.028057295829057693,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.1825037002563477,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.034786805510520935,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.200389862060547,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.03751441463828087,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.2047512531280518,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.035590436309576035,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.1735124588012695,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.03426426649093628,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.199483871459961,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.037030890583992004,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.213322639465332,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.032468877732753754,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.179354190826416,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.027202758938074112,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.216036319732666,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.031184563413262367,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.208205223083496,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.029414812102913857,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.180370569229126,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.02679901383817196,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.210325002670288,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.02863631211221218,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.1601076126098633,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.029135366901755333,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.2262887954711914,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.030535712838172913,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.23160719871521,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.030563877895474434,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.255211353302002,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.02906505949795246,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.2006797790527344,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.030871668830513954,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.2442867755889893,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.02913421392440796,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.218442440032959,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.030087953433394432,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.2164535522460938,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.034147944301366806,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.218660831451416,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.03593966364860535,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.2240614891052246,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.032090652734041214,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.2309441566467285,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.02776424027979374,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.2702016830444336,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.027174271643161774,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.19810152053833,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.028479408472776413,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.2178421020507812,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.026820644736289978,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.2634482383728027,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.026427917182445526,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.2452914714813232,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.02679610252380371,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.243825912475586,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.026115626096725464,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.218787670135498,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.026686571538448334,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.2785303592681885,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.02685859426856041,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.2572226524353027,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.026968171820044518,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.2625553607940674,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.025397246703505516,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.262303352355957,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.025660471990704536,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.269401788711548,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.02713891491293907,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.2416529655456543,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.027591092512011528,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.2814316749572754,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.024462848901748657,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.249725103378296,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.024868478998541832,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.231574535369873,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.027015401050448418,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.282627582550049,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.02678140625357628,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.2208900451660156,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.026830527931451797,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.260575294494629,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.02430890128016472,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.2522997856140137,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.027047613635659218,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.2721078395843506,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.026861369609832764,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.2235946655273438,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.026650549843907356,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.302743434906006,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.02551819011569023,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.265763282775879,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.024843450635671616,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.272397518157959,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.024423206225037575,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.253495454788208,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.024180661886930466,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.260438919067383,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.023554982617497444,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.2342851161956787,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.023470086976885796,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.251622438430786,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.024405231699347496,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.2588815689086914,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.024147814139723778,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.2723565101623535,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.023217910900712013,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.2555792331695557,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.024921158328652382,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.26884126663208,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.02455919235944748,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.2674927711486816,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.024725932627916336,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.2511110305786133,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.025769414380192757,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.290651321411133,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.025928806513547897,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.302396297454834,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.02499072253704071,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.2813124656677246,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.025431277230381966,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.234684705734253,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.026871860027313232,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.2800331115722656,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.02578776702284813,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.3117146492004395,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.026318378746509552,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.257615089416504,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.025542527437210083,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.2969868183135986,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.025723429396748543,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.290259838104248,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.024588774889707565,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.284090280532837,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.024619122967123985,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.284879207611084,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.024441618472337723,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.283104658126831,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.02402760274708271,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.3013617992401123,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.025028521195054054,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.2892568111419678,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.029445001855492592,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 3.2474446296691895,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.128176689147949,
+ "eval_runtime": 58.8588,
+ "eval_samples_per_second": 41.489,
+ "eval_steps_per_second": 1.308,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.0298826452344656,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.138103485107422,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.0387548953294754,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.161492347717285,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.03816070780158043,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.1739673614501953,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.03409738838672638,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.182478189468384,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.033135563135147095,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.1720199584960938,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.03203624114394188,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.1699862480163574,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.031237754970788956,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.179363250732422,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.03342146426439285,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.1822280883789062,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.03561089560389519,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.1767094135284424,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.034710463136434555,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.218817949295044,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.030870331451296806,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.168602705001831,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.028004391118884087,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.224155902862549,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.030953701585531235,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.1789557933807373,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.030289432033896446,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.1890387535095215,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.030684132128953934,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.195427417755127,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.03022196516394615,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.1984965801239014,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.026935536414384842,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.2014272212982178,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.02702922560274601,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.205538034439087,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.026682600378990173,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.1954591274261475,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.0273384228348732,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.1782338619232178,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.028587577864527702,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.199565887451172,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.028848132118582726,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.181786298751831,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.027535341680049896,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.2013628482818604,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.02756025455892086,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.1749587059020996,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.027358995750546455,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.161881923675537,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.025918981060385704,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.225224256515503,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.026479976251721382,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.2129316329956055,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.02685079537332058,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.1651554107666016,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.027378706261515617,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.193108558654785,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.027240974828600883,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.190605640411377,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.028065402060747147,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.2068259716033936,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.028329014778137207,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.2316060066223145,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.028920065611600876,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.1933932304382324,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.028299354016780853,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.192711353302002,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.025996020063757896,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.243171453475952,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.0264595914632082,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.199758529663086,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.02704528719186783,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.2390713691711426,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.026571933180093765,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.219440460205078,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.025562983006238937,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.2538228034973145,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.024683184921741486,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.2372591495513916,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.025656528770923615,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.250558853149414,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.024807540699839592,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.1856563091278076,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.026149194687604904,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.172186851501465,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.02617705799639225,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.250445604324341,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.026542512699961662,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.220273494720459,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.024963509291410446,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.2330946922302246,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.027144264429807663,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.1860971450805664,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.02665683813393116,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.257209062576294,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.024872658774256706,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.2266182899475098,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.025235185399651527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.2134504318237305,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.025029495358467102,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.2150087356567383,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.025630896911025047,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.282879114151001,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.02722341939806938,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.2366206645965576,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.02522372268140316,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.2840523719787598,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.02494790591299534,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.2405319213867188,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.026662737131118774,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.239692211151123,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.026082608848810196,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.2156195640563965,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.02423974685370922,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.1863760948181152,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.025226688012480736,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.20874285697937,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.02428225241601467,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.2209601402282715,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.023819707334041595,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.2564258575439453,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.025661710649728775,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.244133949279785,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.02423049882054329,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.2376630306243896,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.02382473088800907,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.2951290607452393,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.02488875202834606,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.2848310470581055,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.02575223706662655,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.2574503421783447,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.02462848648428917,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.2464442253112793,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.024912388995289803,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.2459192276000977,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.025020884349942207,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.213622808456421,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.026330886408686638,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.2305431365966797,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.02708602324128151,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.2487688064575195,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.030908891931176186,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 3.2110159397125244,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.146673202514648,
+ "eval_runtime": 58.7379,
+ "eval_samples_per_second": 41.574,
+ "eval_steps_per_second": 1.311,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.02919941022992134,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.1533267498016357,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.03379782289266586,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.1436538696289062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.033698830753564835,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.15297532081604,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.03077736124396324,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.128848075866699,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.031112324446439743,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.1461119651794434,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.03346852958202362,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.113943099975586,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.03307225927710533,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.147498607635498,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.030230754986405373,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.1162924766540527,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.02982981689274311,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.144481658935547,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.031735632568597794,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.133862018585205,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.0312766470015049,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.110287666320801,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.029902562499046326,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.1384904384613037,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.0306564774364233,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.1138200759887695,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.03084711916744709,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.155240535736084,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.03183220326900482,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.1593515872955322,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.031826410442590714,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.1517653465270996,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.030141707509756088,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.1291589736938477,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.03158571943640709,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.178101062774658,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.03510458394885063,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.138767719268799,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.03660859912633896,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.1708762645721436,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.034004807472229004,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.1404073238372803,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.03077496401965618,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.159081220626831,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.030384650453925133,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.1872847080230713,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.02957277186214924,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.165973424911499,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.029529055580496788,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.123840808868408,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.027776138857007027,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.1705269813537598,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.028943249955773354,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.1181654930114746,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.028178401291370392,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.201678991317749,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.02865665778517723,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.18479061126709,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.029173649847507477,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.1844916343688965,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.0280170775949955,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.187105417251587,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.029103878885507584,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.183281660079956,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.029045486822724342,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.1848196983337402,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.028364380821585655,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.197348117828369,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.029651399701833725,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.1429920196533203,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.0281438659876585,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.1916513442993164,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.029437677934765816,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.172391176223755,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.03433044254779816,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.2296864986419678,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.03259492665529251,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.1962363719940186,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.0299630519002676,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.169673442840576,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.028495769947767258,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.1855409145355225,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.0290420800447464,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.23667573928833,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.03340664133429527,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.1873788833618164,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.03253626823425293,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.204796552658081,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.028142794966697693,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.193922281265259,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.028805622830986977,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.1863901615142822,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.030033612623810768,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.198482036590576,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.028154220432043076,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.215069055557251,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.028098132461309433,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.2197628021240234,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.028679659590125084,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.2062196731567383,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.029104258865118027,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.2182908058166504,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.029629778116941452,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.2165639400482178,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.028304366394877434,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.1783785820007324,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.026900488883256912,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.1876602172851562,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.027047287672758102,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.1879465579986572,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.027424108237028122,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.2407851219177246,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.029319508001208305,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.1980130672454834,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.027114372700452805,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.211357593536377,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.02579840086400509,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.2067861557006836,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.02767331339418888,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.198587417602539,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.026645265519618988,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.177441120147705,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.026181906461715698,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.226815700531006,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.027218692004680634,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.1731510162353516,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.027551919221878052,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.2268621921539307,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.025782009586691856,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.2052621841430664,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.024231620132923126,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.201770544052124,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.02474607340991497,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.1935834884643555,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.026299327611923218,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.223275899887085,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.02439914271235466,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.2513954639434814,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.02650653012096882,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.1914851665496826,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.024847298860549927,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.2173104286193848,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.028475770726799965,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 3.2184863090515137,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.168707847595215,
+ "eval_runtime": 57.4035,
+ "eval_samples_per_second": 42.541,
+ "eval_steps_per_second": 1.341,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.02837727963924408,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.119239330291748,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.03562312573194504,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.092813014984131,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.0344194732606411,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.0802855491638184,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.03188958019018173,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.1177961826324463,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.032181013375520706,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.1343281269073486,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.031085725873708725,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.118828296661377,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.031995322555303574,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.110109329223633,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.02949182130396366,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.0972743034362793,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.029299668967723846,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.1257238388061523,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.030558617785573006,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.1216022968292236,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.03181575983762741,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.093721866607666,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.030934303998947144,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.0910396575927734,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.030235569924116135,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.133371114730835,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.03350163996219635,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.1178882122039795,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.03468719869852066,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.1222469806671143,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.034203238785266876,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.12453031539917,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.03321516141295433,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.115428924560547,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.03298434242606163,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.127288341522217,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.029987169429659843,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.1194515228271484,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.031271349638700485,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.1284165382385254,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.032201699912548065,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.1912736892700195,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.0299740731716156,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.1233785152435303,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.029613154008984566,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.1535067558288574,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.03303765505552292,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.1550910472869873,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.032644059509038925,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.1367573738098145,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.031695179641246796,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.1350858211517334,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.030685821548104286,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.121593952178955,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.03091946430504322,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.113060235977173,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.0282905176281929,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.184551239013672,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.030020160600543022,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.1673004627227783,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.031114062294363976,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.162588596343994,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.03007666766643524,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.132836103439331,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.029013793915510178,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.1552720069885254,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.02784561738371849,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.189272165298462,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.02759280428290367,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.141814708709717,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.02699868381023407,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.1127257347106934,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.02782190404832363,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.193972587585449,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.027298318222165108,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.1393446922302246,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.027321161702275276,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.0926666259765625,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.026399722322821617,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.1578259468078613,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.02770969271659851,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.1816914081573486,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.028344258666038513,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.1860158443450928,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.028091782703995705,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.1585092544555664,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.026564039289951324,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.131692409515381,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.026349127292633057,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.160431385040283,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.030081138014793396,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.1647543907165527,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.029250919818878174,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.1522226333618164,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.0282477755099535,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.139249324798584,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.0284534003585577,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.154621124267578,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.02677394263446331,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.1375081539154053,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.026511570438742638,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.1355676651000977,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.02766270563006401,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.143874168395996,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.027013977989554405,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.1602487564086914,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.026629449799656868,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.1549456119537354,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.0272684209048748,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.1739020347595215,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.027645915746688843,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.172959327697754,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.025413552299141884,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.1649041175842285,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.028406929224729538,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.1821887493133545,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.027753977105021477,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.163238048553467,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.02693209983408451,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.1921019554138184,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.02671878971159458,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.159438133239746,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.027099428698420525,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.1806793212890625,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.025098824873566628,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.158310651779175,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.02655479498207569,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.1927309036254883,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.025836164131760597,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.1578540802001953,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.02657160349190235,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.1658592224121094,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.027223270386457443,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.1760241985321045,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.025695348158478737,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.202389717102051,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.025819813832640648,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.2111103534698486,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.02621106244623661,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.1560299396514893,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.026585502550005913,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.176602840423584,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.03038652054965496,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 3.1799943447113037,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.193845748901367,
+ "eval_runtime": 57.9911,
+ "eval_samples_per_second": 42.11,
+ "eval_steps_per_second": 1.328,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.03135736659169197,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.070986270904541,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.038986921310424805,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.0708835124969482,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.035115309059619904,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.1070001125335693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.032868221402168274,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.082258701324463,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.03703184798359871,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.057744264602661,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.03831101953983307,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.0789475440979004,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.036301176995038986,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.0565457344055176,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.03166409209370613,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.0574910640716553,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.02984458953142166,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.080018997192383,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.034001126885414124,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.0813302993774414,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.03814122825860977,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.0579185485839844,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.03870106488466263,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.0857529640197754,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.04005642607808113,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.0880727767944336,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.03760755434632301,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.0795767307281494,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.03260907158255577,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.1006722450256348,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.029866183176636696,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.0850343704223633,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.030988436192274094,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.0743892192840576,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.03698134422302246,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.056002378463745,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.03696191683411598,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.069124698638916,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.03453022986650467,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.1017374992370605,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.032458193600177765,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.1166419982910156,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.02892368845641613,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.0863287448883057,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.030625877901911736,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.0938801765441895,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.030280394479632378,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.086432933807373,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.029181072488427162,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.0977859497070312,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.028039127588272095,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.112367868423462,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.02910272404551506,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.1253533363342285,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.03086872771382332,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.103689670562744,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.02883610874414444,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.111599922180176,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.028490116819739342,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.1410303115844727,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.028714407235383987,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.133023738861084,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.029553238302469254,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.107003688812256,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.02997349575161934,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.135680675506592,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.029087591916322708,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.1103882789611816,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.029947761446237564,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.0961623191833496,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.02950555644929409,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.1162495613098145,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.029569562524557114,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.096372127532959,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.031366728246212006,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.120089530944824,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.0313473716378212,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.100785732269287,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.028021851554512978,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.0940558910369873,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.027477117255330086,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.1210484504699707,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.029690749943256378,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.1350059509277344,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.029659733176231384,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.1220788955688477,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.028260033577680588,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.1000189781188965,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.027554461732506752,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.1044349670410156,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.027929967269301414,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.1353745460510254,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.029309319332242012,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.1578292846679688,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.029183287173509598,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.1217150688171387,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.02707512304186821,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.119617223739624,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.03006536327302456,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.1481480598449707,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.03026784397661686,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.142770767211914,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.02827967144548893,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.1582541465759277,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.028892774134874344,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.1530709266662598,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.03023783676326275,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.1102521419525146,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.027951840311288834,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.139113426208496,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.02807226963341236,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.149428367614746,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.029721451923251152,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.120518207550049,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.02867680788040161,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.1109931468963623,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.02663845382630825,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.116053581237793,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.028366081416606903,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.1444268226623535,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.028931576758623123,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.163870096206665,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.02664567157626152,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.1349146366119385,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.028031354770064354,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.1638219356536865,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.029421644285321236,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.1601340770721436,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.027983693405985832,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.1636788845062256,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.02719871513545513,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.1246869564056396,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.03030397929251194,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.1749579906463623,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.029340893030166626,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.1751790046691895,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.027624106034636497,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.178704023361206,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.02830546908080578,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.177117347717285,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.03168068826198578,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.125427722930908,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.033327918499708176,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 3.13071870803833,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.218388080596924,
+ "eval_runtime": 57.611,
+ "eval_samples_per_second": 42.388,
+ "eval_steps_per_second": 1.337,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.03336749225854874,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.064338445663452,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.038479577749967575,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.0478687286376953,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.032562077045440674,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.019803762435913,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.0331728458404541,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.0599141120910645,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.03361169621348381,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.0315823554992676,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.032973092049360275,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.025928258895874,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.030388766899704933,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.070671558380127,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.030899515375494957,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.012938976287842,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.0335412323474884,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.0156331062316895,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.035731345415115356,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.0578935146331787,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.03379053995013237,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.048701286315918,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.031242679804563522,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.0832977294921875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.030997611582279205,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.0663583278656006,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.03397331386804581,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.0115303993225098,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.03349050134420395,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.02864933013916,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.031079065054655075,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.053401470184326,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.029346197843551636,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.0700926780700684,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.030784208327531815,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.066525936126709,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.03368140384554863,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.0956242084503174,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.03356792405247688,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.0394320487976074,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.029909325763583183,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.031525135040283,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.030711578205227852,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.075474977493286,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.03452072665095329,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.0228147506713867,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.03235075995326042,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.0664501190185547,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.029671598225831985,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.055302143096924,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.033868469297885895,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.053016185760498,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.0329158715903759,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.1017332077026367,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.031983695924282074,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.0706472396850586,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.02963697351515293,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.0711426734924316,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.03162391856312752,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.044867515563965,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.030773654580116272,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.0963196754455566,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.031220417469739914,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.0892245769500732,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.030928561463952065,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.1181323528289795,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.03192073106765747,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.066584348678589,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.03075495734810829,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.0436511039733887,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.02895325981080532,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.123373031616211,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.029738714918494225,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.1049060821533203,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.031937118619680405,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.099111318588257,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.029886579141020775,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.0970301628112793,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.029569782316684723,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.062499761581421,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.02924283780157566,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.0618538856506348,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.030955800786614418,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.0745606422424316,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.03145027533173561,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.122938632965088,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.028926467522978783,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.0808773040771484,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.029592076316475868,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.1167049407958984,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.028070243075489998,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.0781636238098145,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.028642063960433006,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.093937397003174,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.0290757454931736,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.1079559326171875,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.027867283672094345,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.0885753631591797,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.02896495908498764,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.0889246463775635,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.02777806483209133,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.084439277648926,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.02904277853667736,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.125246047973633,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.029010627418756485,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.111422061920166,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.028178894892334938,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.115067481994629,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.02763799950480461,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.1228506565093994,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.028052521869540215,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.1150546073913574,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.029438253492116928,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.0921788215637207,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.028195543214678764,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.1408450603485107,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.02937602624297142,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.111612319946289,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.028430236503481865,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.093441963195801,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.027624569833278656,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.1184372901916504,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.029797039926052094,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.1074559688568115,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.028728218749165535,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.161461114883423,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.02704412117600441,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.13325834274292,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.028067929670214653,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.093172550201416,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.028848597779870033,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.1461825370788574,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.02956785447895527,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.141045570373535,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.028214773163199425,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.136033773422241,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.029155246913433075,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.0986382961273193,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.02968468703329563,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.11630916595459,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.028356682509183884,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.1134822368621826,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.03301718086004257,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 3.118591785430908,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.246211528778076,
+ "eval_runtime": 57.5155,
+ "eval_samples_per_second": 42.458,
+ "eval_steps_per_second": 1.339,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.033637888729572296,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.0192506313323975,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.04057888314127922,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 2.977440595626831,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.03579854592680931,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 2.9902219772338867,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.03131850063800812,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 2.9988656044006348,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.03262636810541153,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.037747621536255,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.03260403499007225,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 2.987508773803711,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.033165134489536285,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.0511183738708496,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.03303489834070206,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.0083041191101074,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.03185006603598595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 2.9966228008270264,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.03135651350021362,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.036759376525879,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.03216925635933876,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.0304150581359863,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.03231671452522278,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.0300474166870117,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.03263348713517189,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.0298163890838623,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.032594650983810425,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 2.98789119720459,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.03406650573015213,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.0102760791778564,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.03379780799150467,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.0564732551574707,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.03172587603330612,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.0410943031311035,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.02938298135995865,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.0517544746398926,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.03206252679228783,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.0352776050567627,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.033386752009391785,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.020535945892334,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.033984702080488205,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.0552730560302734,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0329931266605854,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.0075573921203613,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.032101184129714966,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.046388864517212,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.032873060554265976,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.0340287685394287,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.030672110617160797,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.0347838401794434,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.030014054849743843,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.0315332412719727,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.02932834066450596,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.084340810775757,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.029181934893131256,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.032886028289795,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.02966221794486046,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.032766819000244,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.029713116586208344,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.028778076171875,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.029243769124150276,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.0492146015167236,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.02997511625289917,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.0772786140441895,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.02930292673408985,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.055750608444214,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.03037521243095398,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.033710479736328,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.029957240447402,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.0394721031188965,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.02925589308142662,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.0878353118896484,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.03003518097102642,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.048628330230713,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.031545430421829224,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.0775623321533203,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.031310439109802246,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.01645565032959,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.03074057213962078,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.0651440620422363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.028704499825835228,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.0527148246765137,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.029061976820230484,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.0854501724243164,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.03034680150449276,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.056356906890869,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.028715113177895546,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.0874247550964355,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.030256496742367744,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.052424430847168,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.02949833683669567,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.0663650035858154,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.028980428352952003,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.0700113773345947,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.028526026755571365,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.0856711864471436,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.029572727158665657,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.0862035751342773,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.02955584041774273,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.030802011489868,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.02881692908704281,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.074392318725586,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.029564999043941498,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.057894229888916,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.030315004289150238,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.096233367919922,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.030666006729006767,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.040799617767334,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.028901133686304092,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.055051803588867,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.030359387397766113,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.085972309112549,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.03117106668651104,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.0786352157592773,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.029238386079669,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.0375771522521973,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.029324093833565712,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.1140198707580566,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.02875387854874134,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.090881109237671,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.029023943468928337,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.0755701065063477,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.0291463453322649,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.0723776817321777,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.02795177884399891,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.058993339538574,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.029499078169465065,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.0856404304504395,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.029889609664678574,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.0622942447662354,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.03107610158622265,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.114957332611084,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.02967257983982563,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.088871479034424,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.029068559408187866,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.1103081703186035,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.029334546998143196,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.0841622352600098,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.029259610921144485,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.0662848949432373,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.030072513967752457,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.0900328159332275,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.033321622759103775,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 3.0871129035949707,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.271463394165039,
+ "eval_runtime": 57.9048,
+ "eval_samples_per_second": 42.173,
+ "eval_steps_per_second": 1.33,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.03635408729314804,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 2.9576284885406494,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.03927835449576378,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 2.999624729156494,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.03371119126677513,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 2.971052646636963,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.033049341291189194,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 2.980811595916748,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.03830729424953461,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 2.9985947608947754,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.036573827266693115,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 2.9892568588256836,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.031196987256407738,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 2.9854254722595215,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.03376495838165283,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 2.9840264320373535,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.036272983998060226,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.0168871879577637,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.03328242897987366,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 2.9815053939819336,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.03158426657319069,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 2.994525909423828,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.03320910036563873,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.008638858795166,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.03160254284739494,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 2.979243040084839,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.03158991411328316,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 2.938634157180786,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.031116485595703125,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 2.9957656860351562,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.030757686123251915,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.0106635093688965,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.02992827817797661,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 2.996793746948242,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.0330309197306633,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 2.9859161376953125,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.029488038271665573,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.0081536769866943,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.03095114603638649,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.006385087966919,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.03194110840559006,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.0125250816345215,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.03306785598397255,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 2.9640283584594727,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.030883541330695152,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.0223653316497803,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.03224025294184685,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.0274910926818848,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.03182433918118477,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.047736406326294,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.03387221321463585,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 2.9821343421936035,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.03407924994826317,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.041207790374756,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.031546324491500854,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.0298609733581543,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.03354770317673683,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.0212411880493164,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.031692177057266235,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.007350444793701,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.03218817710876465,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.0019829273223877,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.03233771771192551,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.0400614738464355,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.028998777270317078,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.022221088409424,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.030422436073422432,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.0071797370910645,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.030207227915525436,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 2.9762964248657227,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.0305719543248415,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.075176239013672,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.03173545375466347,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.028982162475586,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.02799576334655285,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.0242385864257812,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.03140150383114815,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.012221336364746,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.029416466131806374,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.0037949085235596,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.028385788202285767,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.0350992679595947,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.031597770750522614,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.065139055252075,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.02926146425306797,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.014230728149414,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.02970231883227825,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.06636381149292,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.029684338718652725,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.024610996246338,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.02908361330628395,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.0093719959259033,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.029891187325119972,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.0735411643981934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.028607673943042755,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.044274091720581,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.029806701466441154,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.0449328422546387,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.03161340951919556,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.046813488006592,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.028860243037343025,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.011828660964966,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.0285879448056221,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.043640613555908,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.0296401996165514,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.071589469909668,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.02807026542723179,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.018320083618164,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.029037052765488625,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.0471763610839844,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.029226379469037056,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.037463426589966,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.029322804883122444,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.0472755432128906,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.028253905475139618,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.0265414714813232,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.028875533491373062,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.063941717147827,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.028460513800382614,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.0512683391571045,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.02951696142554283,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.0528573989868164,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.030593490228056908,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.057919979095459,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.028074532747268677,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.0263280868530273,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.02955678664147854,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.081322193145752,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.029724644497036934,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.059583902359009,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.029713226482272148,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.018678903579712,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.029758386313915253,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.0379555225372314,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.028971998021006584,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.041267156600952,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.02968725562095642,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.0675506591796875,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.029876509681344032,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.051586151123047,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.02853119559586048,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.0335168838500977,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.03374241292476654,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 3.027909278869629,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.030695006251335144,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 2.9685583114624023,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0339728482067585,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 2.936551570892334,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.03372679278254509,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 2.9631872177124023,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.03172554448246956,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 2.969510078430176,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.030516354367136955,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 2.930342674255371,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.03261839970946312,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 2.9920477867126465,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.030618635937571526,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 2.9693596363067627,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.030576011165976524,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 2.929804801940918,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.03224785998463631,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 2.95280122756958,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.030605586245656013,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 2.931358575820923,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.031206639483571053,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 2.962998151779175,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.031317099928855896,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 2.9706358909606934,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.031614501029253006,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 2.9326977729797363,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.031002728268504143,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 2.99625301361084,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.03216253221035004,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 2.920912027359009,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.033380527049303055,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 2.9688234329223633,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.03085758350789547,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 2.9811222553253174,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.03258587792515755,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 2.932366371154785,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.031639691442251205,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 2.988865613937378,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.03144839406013489,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 2.9750723838806152,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.030864672735333443,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 2.999329090118408,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.030679022893309593,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 2.974264144897461,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.03236627206206322,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 2.9808404445648193,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.03115275874733925,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.0216290950775146,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.029868317767977715,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 2.9739654064178467,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.03446527197957039,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 2.9565515518188477,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.033536117523908615,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 2.958047389984131,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.03113192319869995,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 2.9662246704101562,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.03089931420981884,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 2.9972119331359863,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.030822310596704483,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.020500421524048,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.032264843583106995,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 2.9895882606506348,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.031226349994540215,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 2.958217144012451,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.03110436722636223,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 2.9884791374206543,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.03358910232782364,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 2.9908103942871094,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.03348501771688461,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 2.985677480697632,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.029595404863357544,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 2.979318380355835,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.031459368765354156,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 2.9921798706054688,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.03067564032971859,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 2.977778434753418,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.030506785959005356,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 2.9841768741607666,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.029978664591908455,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 2.995443344116211,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.030900927260518074,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.0177016258239746,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.029710259288549423,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.01532244682312,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.031190430745482445,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 2.9791834354400635,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.03150483965873718,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 2.998800754547119,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.030249621719121933,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.0187838077545166,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.031052686274051666,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 2.984062433242798,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.02961024083197117,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 2.970818042755127,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.03020382858812809,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 2.9913196563720703,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.031914591789245605,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 2.993145704269409,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.03061537817120552,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.0131571292877197,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.029960963875055313,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.0017619132995605,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.030348235741257668,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.0220556259155273,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.03030729480087757,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.0499038696289062,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.030809229239821434,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 2.986651659011841,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.029438305646181107,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.0351314544677734,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.029887216165661812,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.021660327911377,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.02955588512122631,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 2.9990694522857666,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.029586469754576683,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.0296952724456787,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.030632557347416878,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.0124430656433105,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.030579863116145134,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.0381762981414795,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.03282996267080307,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.0329041481018066,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.031176412478089333,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.037598133087158,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.028640294447541237,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.0004701614379883,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.0326533280313015,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.0328853130340576,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.03041139990091324,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.0149893760681152,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.030366195365786552,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.0379695892333984,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.03106893040239811,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.0284576416015625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.0288887657225132,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.028569221496582,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.03154266998171806,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.0737850666046143,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.0305064395070076,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.0744194984436035,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.030967431142926216,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.0170607566833496,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.036945272237062454,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 3.0355610847473145,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.329805374145508,
+ "eval_runtime": 64.1339,
+ "eval_samples_per_second": 38.077,
+ "eval_steps_per_second": 1.201,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.03236053138971329,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 2.9230728149414062,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.0395631417632103,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 2.92578125,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.0361955352127552,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 2.937826633453369,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.035918429493904114,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 2.9481167793273926,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.03309318795800209,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 2.8919506072998047,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0327315591275692,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 2.903585433959961,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.033141329884529114,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 2.942802667617798,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.033966805785894394,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 2.9210848808288574,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.03218021243810654,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 2.933588981628418,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.03272828459739685,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 2.9202046394348145,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.030764421448111534,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 2.916121006011963,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.031897611916065216,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 2.957679033279419,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.03199796751141548,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 2.9276621341705322,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.030233927071094513,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 2.9006078243255615,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.03035062551498413,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 2.920341730117798,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.03029404580593109,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 2.937875747680664,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.029371997341513634,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 2.9832730293273926,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.030163884162902832,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 2.94118332862854,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.03072655200958252,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 2.9632580280303955,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.02975962497293949,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 2.9346470832824707,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.029814405366778374,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 2.9451403617858887,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.030135225504636765,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 2.944918155670166,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.029395751655101776,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 2.925586223602295,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.0306396484375,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 2.9487359523773193,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.03110465221107006,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 2.963491678237915,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.03155078738927841,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 2.9355576038360596,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.030103784054517746,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 2.953045606613159,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.030827803537249565,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 2.968574047088623,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.0310911163687706,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 2.977902412414551,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.030260782688856125,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 2.94873309135437,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.030654428526759148,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 2.9282753467559814,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.02993479184806347,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 2.97076416015625,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.030244970694184303,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 2.960441827774048,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.030003050342202187,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 2.9672164916992188,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.030057862401008606,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 2.961918830871582,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.029379919171333313,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 2.9918394088745117,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.03060895763337612,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 2.94891619682312,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.030228130519390106,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 2.9726617336273193,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.029833678156137466,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 2.9494335651397705,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.030842263251543045,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 2.9580483436584473,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.029707176610827446,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 2.963684320449829,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.030966786667704582,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 2.9710958003997803,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.03137032687664032,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.003589391708374,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.03123190440237522,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 2.9742860794067383,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.030608588829636574,
+ "learning_rate": 0.000110418175419276,
+ "loss": 2.97404408454895,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.030018644407391548,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 2.972231864929199,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.03056846372783184,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 2.9871716499328613,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.02888435684144497,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 2.9774169921875,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.03173990175127983,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 2.941751480102539,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.030047327280044556,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 2.990516185760498,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.030729815363883972,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 2.9745054244995117,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.030600983649492264,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 2.9676642417907715,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.03208782523870468,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.006941556930542,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.031162556260824203,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 2.9740395545959473,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.031870320439338684,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 2.9809296131134033,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.03184300288558006,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 2.9781875610351562,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.029629364609718323,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 2.9912891387939453,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.03322628512978554,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 2.9957847595214844,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.0322413295507431,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 2.9894957542419434,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.030753307044506073,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.0320589542388916,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.03447900339961052,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.0189106464385986,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.032016605138778687,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.0162312984466553,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.030838409438729286,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.000263214111328,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.034556787461042404,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 2.996680736541748,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.03266173601150513,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 2.990510940551758,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.030973197892308235,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 2.9767255783081055,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.03442178666591644,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 2.9941515922546387,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.03215927258133888,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.006618022918701,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.030949831008911133,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.004136085510254,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.03256693854928017,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.0229580402374268,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.02986278384923935,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 2.9580869674682617,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.03597614914178848,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.995757579803467,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.359405517578125,
+ "eval_runtime": 68.1721,
+ "eval_samples_per_second": 35.821,
+ "eval_steps_per_second": 1.129,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.03432514891028404,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 2.922243118286133,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.0345693901181221,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 2.9114043712615967,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0349399633705616,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 2.8668057918548584,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.03161955252289772,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 2.8996763229370117,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.032124556601047516,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 2.8662075996398926,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.03291172534227371,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 2.9088220596313477,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.033552397042512894,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 2.9076967239379883,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.03218761831521988,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 2.8726117610931396,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.03289451450109482,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 2.9127345085144043,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.033019524067640305,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 2.920897960662842,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.03153533488512039,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 2.9124374389648438,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.032293178141117096,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 2.877277374267578,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.031969670206308365,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 2.921635150909424,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.03023488074541092,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 2.9336061477661133,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.03215106576681137,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 2.905877113342285,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.030415233224630356,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 2.909331798553467,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.031150560826063156,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 2.9302995204925537,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.03152180463075638,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 2.9171547889709473,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.02996763028204441,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 2.916459798812866,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.030544105917215347,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 2.898505449295044,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.030963370576500893,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 2.948704719543457,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.03191790357232094,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 2.9182960987091064,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.02956460975110531,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 2.9211082458496094,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.031227655708789825,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 2.942269802093506,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.031859587877988815,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 2.9169349670410156,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.03045276179909706,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 2.9240684509277344,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.031577471643686295,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 2.9135894775390625,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.031031791120767593,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 2.913355827331543,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.030012110248208046,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 2.8860886096954346,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.029772799462080002,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 2.974332809448242,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.029961293563246727,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 2.943694591522217,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.029537590220570564,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 2.93180513381958,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.030834559351205826,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 2.9442358016967773,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.02998523600399494,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 2.9336910247802734,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.030030962079763412,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 2.93593692779541,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.030250679701566696,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 2.9370553493499756,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.02942339889705181,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 2.9406495094299316,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.030448907986283302,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 2.9552698135375977,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.03078317642211914,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 2.933631420135498,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.029188675805926323,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 2.944002628326416,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.03153637796640396,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 2.9708166122436523,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.030299564823508263,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 2.9570717811584473,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.03169458732008934,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 2.942774772644043,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.031527288258075714,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 2.9711742401123047,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.0314360186457634,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 2.9509239196777344,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.032102152705192566,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 2.9443840980529785,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.03111223131418228,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 2.972594738006592,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.0322272814810276,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 2.9361982345581055,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.033640168607234955,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 2.9726009368896484,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.02979348786175251,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 2.9978041648864746,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.030464721843600273,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 2.949810028076172,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.03170578181743622,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 2.9677181243896484,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.02996704913675785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 2.974087953567505,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.030277637764811516,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 2.95644211769104,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.030859284102916718,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 2.957258701324463,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.029737796634435654,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 2.9502854347229004,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.029822375625371933,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 2.9625749588012695,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.030455606058239937,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 2.9669153690338135,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.029491793364286423,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 2.952991247177124,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.030326122418045998,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 2.978355646133423,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.03011198714375496,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 2.976996421813965,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.029053322970867157,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 2.9754371643066406,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.02925761416554451,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 2.94271183013916,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.029678862541913986,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 2.9386143684387207,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.031623054295778275,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 2.9660329818725586,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.030133802443742752,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 2.9580249786376953,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.029482144862413406,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 2.944474697113037,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.030708568170666695,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 2.967862367630005,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.028793595731258392,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 2.9581642150878906,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.03083081543445587,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 2.957261323928833,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.030180038884282112,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 2.9579484462738037,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.034881673753261566,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.9007070064544678,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.38953161239624,
+ "eval_runtime": 71.4964,
+ "eval_samples_per_second": 34.156,
+ "eval_steps_per_second": 1.077,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0335528701543808,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 2.8542537689208984,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.03163984417915344,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 2.864593505859375,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.034005917608737946,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 2.855072498321533,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.03235425800085068,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 2.8504371643066406,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.03171464428305626,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 2.8924291133880615,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.0333196297287941,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 2.8764119148254395,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.03050111047923565,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 2.874119281768799,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.032114267349243164,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 2.9059693813323975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.03141588345170021,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 2.879429340362549,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.03201865404844284,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 2.898519277572632,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.03152940049767494,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 2.866131544113159,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.032520636916160583,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 2.888209819793701,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.03244909271597862,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 2.888913631439209,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.03160687908530235,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 2.905120372772217,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.031518299132585526,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 2.8985815048217773,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.0304370429366827,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 2.8531291484832764,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.03239843621850014,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 2.876098394393921,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.03194732964038849,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 2.908271074295044,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.03145617991685867,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 2.874128580093384,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.031182898208498955,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 2.894904613494873,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.0323818176984787,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 2.9005861282348633,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.030969973653554916,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 2.9236338138580322,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.03238993510603905,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 2.8811020851135254,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.030876649543642998,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 2.8802058696746826,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.032251305878162384,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 2.882112503051758,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.03334304690361023,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 2.9290661811828613,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.03299225866794586,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 2.908268928527832,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.03335456922650337,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 2.9156572818756104,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.030786098912358284,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 2.8981709480285645,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.03379470854997635,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 2.9483864307403564,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.03020016849040985,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 2.908958911895752,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.032774876803159714,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 2.899505138397217,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.030589981004595757,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 2.9257779121398926,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.03366618603467941,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 2.8749284744262695,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.03038695827126503,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 2.9127252101898193,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.03400666266679764,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 2.922375202178955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.031702589243650436,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 2.8886446952819824,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.03223840892314911,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 2.906278610229492,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.0327419750392437,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 2.9436042308807373,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.030698910355567932,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 2.897615909576416,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.03250030800700188,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 2.928028106689453,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.030623450875282288,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 2.9268975257873535,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.030981825664639473,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 2.938185691833496,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.03139727562665939,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 2.9397618770599365,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.029906271025538445,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 2.921111583709717,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.03110392577946186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 2.899641513824463,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.03060143068432808,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 2.922635078430176,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.03061371110379696,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 2.925189733505249,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.03080526366829872,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 2.90329647064209,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.030412547290325165,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 2.9444844722747803,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.03024943359196186,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 2.9165825843811035,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.030653616413474083,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 2.9171371459960938,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.03027702495455742,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 2.9399702548980713,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.030466679483652115,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 2.888066291809082,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.02997267246246338,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 2.906545877456665,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.03026283159852028,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 2.9359335899353027,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.0293166171759367,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 2.9118804931640625,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.031117571517825127,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 2.968198299407959,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.03041299246251583,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 2.9333484172821045,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.03037327155470848,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 2.895859956741333,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.03161945194005966,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 2.9491944313049316,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.031058384105563164,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 2.926038980484009,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.030037537217140198,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 2.941394805908203,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.030748063698410988,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 2.998904228210449,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.030324939638376236,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 2.9226560592651367,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.030745431780815125,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 2.977520227432251,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.02952338382601738,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 2.9365828037261963,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.030132673680782318,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 2.9574036598205566,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.030379803851246834,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 2.9329113960266113,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.029670827090740204,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 2.951117515563965,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.029823293909430504,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 2.9586987495422363,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.03546741604804993,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.9378662109375,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.417913436889648,
+ "eval_runtime": 69.6157,
+ "eval_samples_per_second": 35.078,
+ "eval_steps_per_second": 1.106,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.033859945833683014,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 2.8493661880493164,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.032931916415691376,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 2.844602108001709,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.03232948109507561,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 2.8463873863220215,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.032815027981996536,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 2.846658706665039,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.03269147500395775,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 2.8286149501800537,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.03337012976408005,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 2.9202351570129395,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.031733594834804535,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 2.8601999282836914,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.032877493649721146,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 2.8566389083862305,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.033852867782115936,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 2.859797477722168,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.03143248334527016,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 2.8411474227905273,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.033198583871126175,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 2.888702392578125,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.029946891590952873,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 2.8332901000976562,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.031660862267017365,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 2.8255481719970703,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.030189910903573036,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 2.8664190769195557,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.03273668885231018,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 2.886531114578247,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.030584124848246574,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 2.872990131378174,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.032543931156396866,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 2.891528606414795,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.031016048043966293,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 2.8759872913360596,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.03190434351563454,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 2.8757526874542236,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.030996622517704964,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 2.862654447555542,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.03289523348212242,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 2.9064857959747314,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.03322640061378479,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 2.8921048641204834,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.0303682629019022,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 2.886458396911621,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.033412594348192215,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 2.8638253211975098,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.030457086861133575,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 2.835578441619873,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.03225106745958328,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 2.8652632236480713,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.03178466483950615,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 2.8796353340148926,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.03075973503291607,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 2.895150661468506,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.02984696812927723,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 2.882878303527832,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.030945835635066032,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 2.898838520050049,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.030904723331332207,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 2.885678291320801,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.030709490180015564,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 2.8645198345184326,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.030528701841831207,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 2.8877408504486084,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.030471473932266235,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 2.9136905670166016,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.030309896916151047,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 2.9238810539245605,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.030605493113398552,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 2.8624460697174072,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.029686832800507545,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 2.9085543155670166,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.030039243400096893,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 2.8624000549316406,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.02953776717185974,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 2.8893938064575195,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.029651308432221413,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 2.914975166320801,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.030902806669473648,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 2.887071132659912,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.02948773466050625,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 2.8893041610717773,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.03049607202410698,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 2.902736186981201,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.030077926814556122,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 2.8936967849731445,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.0303986556828022,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 2.9223172664642334,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.030650658532977104,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 2.9140920639038086,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.029562659561634064,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 2.918179988861084,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.03063463792204857,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 2.9222826957702637,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.030284354463219643,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 2.8685007095336914,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.029791876673698425,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 2.914102554321289,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.030152756720781326,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 2.8890669345855713,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.02957645244896412,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 2.8851962089538574,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.02963421493768692,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 2.9012060165405273,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.029279837384819984,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 2.949984073638916,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.0302849430590868,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 2.9049415588378906,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.029432371258735657,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 2.8919944763183594,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.030542058870196342,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 2.9168801307678223,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.02911970019340515,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 2.9342129230499268,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.030335791409015656,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 2.918611764907837,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.03031272068619728,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 2.8769750595092773,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.030200788751244545,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 2.8994460105895996,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.03032573312520981,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 2.920720100402832,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.03180326521396637,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 2.891145706176758,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.029451677575707436,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 2.9057083129882812,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.030261319130659103,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 2.9132461547851562,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.031257566064596176,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 2.9089136123657227,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.03043205663561821,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 2.912537097930908,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.030708536505699158,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 2.8841137886047363,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.03044111281633377,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 2.921257257461548,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.029576802626252174,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 2.908468723297119,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.03031427040696144,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 2.936445713043213,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.036312974989414215,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.8997855186462402,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.443755149841309,
+ "eval_runtime": 62.9898,
+ "eval_samples_per_second": 38.768,
+ "eval_steps_per_second": 1.222,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.03337511047720909,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 2.8584513664245605,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.031415343284606934,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 2.8529202938079834,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.033063437789678574,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 2.814058780670166,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.03153088316321373,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 2.80460786819458,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.03227711841464043,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 2.858191967010498,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.030423127114772797,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 2.845815420150757,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.032030411064624786,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 2.837918281555176,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.030464086681604385,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 2.841099739074707,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.03194798156619072,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 2.843568801879883,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.03152863308787346,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 2.8779497146606445,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.03201183304190636,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 2.906585216522217,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.02999979443848133,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 2.84859561920166,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.03106156550347805,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 2.814643383026123,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.03118651732802391,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 2.8370001316070557,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.031022636219859123,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 2.8466384410858154,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.03163835406303406,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 2.808440685272217,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.03071477636694908,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 2.850621223449707,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.03041849471628666,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 2.818002939224243,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.03104640357196331,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 2.8375182151794434,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.02970169298350811,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 2.871188163757324,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.03146181255578995,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 2.8356661796569824,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.030204759910702705,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 2.861675262451172,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.029957789927721024,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 2.8493146896362305,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.031224241480231285,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 2.876760959625244,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 0.02914278209209442,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 2.838275909423828,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.03204638510942459,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 2.8679757118225098,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.030257491394877434,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 2.8245105743408203,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.031307291239500046,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 2.852024793624878,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 0.030594198033213615,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 2.883610248565674,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.03119277022778988,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 2.866126775741577,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.0302574560046196,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 2.857886791229248,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.029702702537178993,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 2.8324077129364014,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.029572894796729088,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 2.878951072692871,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.03024783544242382,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 2.900284767150879,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.02952091582119465,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 2.8402292728424072,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.03006756864488125,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 2.8609299659729004,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.029146816581487656,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 2.887960433959961,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.03026733547449112,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 2.852726459503174,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.029839467257261276,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 2.8559489250183105,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.029919864609837532,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 2.880255937576294,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.03074759431183338,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 2.8586673736572266,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.029141079634428024,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 2.8813042640686035,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.031023059040308,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 2.8355565071105957,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.029786985367536545,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 2.872175693511963,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.030190780758857727,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 2.874236583709717,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.029669199138879776,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 2.8723950386047363,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.030073825269937515,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 2.863720417022705,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.02976175583899021,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 2.883267402648926,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.030071711167693138,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 2.8690664768218994,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.031068995594978333,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 2.9107182025909424,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.030393972992897034,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 2.859959602355957,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.030435480177402496,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 2.9003896713256836,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.02982497587800026,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 2.857231378555298,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.030999863520264626,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 2.8567168712615967,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.029698362573981285,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 2.8949484825134277,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.031584806740283966,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 2.8703250885009766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.02952221967279911,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 2.8914244174957275,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.03138599172234535,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 2.917753219604492,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.029678434133529663,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 2.8854236602783203,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.03174183890223503,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 2.8508100509643555,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.03006329946219921,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 2.883819103240967,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.031188689172267914,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 2.932220458984375,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.029729023575782776,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 2.894927978515625,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.02953082136809826,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 2.896371364593506,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.030423447489738464,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 2.9003541469573975,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.029901325702667236,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 2.8726208209991455,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.030071355402469635,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 2.881291151046753,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.029847780242562294,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 2.944478988647461,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.029794953763484955,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 2.8625364303588867,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.03029358573257923,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 2.9106764793395996,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.031095417216420174,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 2.8879761695861816,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.03671329468488693,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.877009391784668,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 4.470407962799072,
+ "eval_runtime": 57.8161,
+ "eval_samples_per_second": 42.237,
+ "eval_steps_per_second": 1.332,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.03383656591176987,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 2.8301892280578613,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.030587317422032356,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 2.809070587158203,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.034895412623882294,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 2.845365285873413,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.03129304572939873,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 2.8176681995391846,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.03369338437914848,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 2.820741653442383,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.030895931646227837,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 2.810272693634033,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.03239824250340462,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 2.855811595916748,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.03123638592660427,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 2.7941160202026367,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.031117001548409462,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 2.784010887145996,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.03162471950054169,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 2.8337666988372803,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.0316263884305954,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 2.8019790649414062,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.031872715801000595,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 2.8156003952026367,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.030252376571297646,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 2.7971153259277344,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.03168163076043129,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 2.791532039642334,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.03014831617474556,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 2.8359432220458984,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.03142879158258438,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 2.846064567565918,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.030883461236953735,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 2.8668103218078613,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.03189866617321968,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 2.8214638233184814,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.029116913676261902,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 2.850940704345703,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.030518673360347748,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 2.844576358795166,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.029227688908576965,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 2.797168493270874,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.029839880764484406,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 2.8534903526306152,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.03011847473680973,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 2.8278157711029053,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.03027244098484516,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 2.858158588409424,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.030171507969498634,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 2.8523292541503906,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.029657812789082527,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 2.8015360832214355,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.028928041458129883,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 2.8357324600219727,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.029674772173166275,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 2.8269171714782715,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.03013371117413044,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 2.822054862976074,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.02960435301065445,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 2.8452866077423096,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.030388740822672844,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 2.855799674987793,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.029823970049619675,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 2.8262312412261963,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.030613111332058907,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 2.8490824699401855,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.029467277228832245,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 2.855149984359741,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.030264565721154213,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 2.849930763244629,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.02961486577987671,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 2.843088150024414,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.029051944613456726,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 2.825998067855835,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.03038879856467247,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 2.8565316200256348,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.02917695790529251,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 2.8733534812927246,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.03011898323893547,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 2.8347158432006836,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.03015534020960331,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 2.878455400466919,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.02992299199104309,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 2.861504554748535,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.030217986553907394,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 2.8451390266418457,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.029564134776592255,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 2.8065595626831055,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.03130771592259407,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 2.8506879806518555,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.030411556363105774,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 2.857304573059082,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.030868662521243095,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 2.8545470237731934,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.03002075105905533,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 2.889376640319824,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.030850842595100403,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 2.8634769916534424,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.03078618459403515,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 2.8487653732299805,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.031029092147946358,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 2.897327184677124,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.03101371042430401,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 2.8642020225524902,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.03044370748102665,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 2.87313175201416,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.030529852956533432,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 2.849147081375122,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.029675021767616272,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 2.8475494384765625,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.03002132847905159,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 2.8713879585266113,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.03048008307814598,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 2.8976492881774902,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.02930760383605957,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 2.850775718688965,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.029840776696801186,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 2.8743128776550293,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.030162667855620384,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 2.850478410720825,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.029736310243606567,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 2.851759195327759,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.029385298490524292,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 2.8473241329193115,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.028572317212820053,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 2.852155923843384,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.0300681721419096,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 2.871021270751953,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.0292969960719347,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 2.8470516204833984,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.029493169859051704,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 2.892655372619629,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.029351307079195976,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 2.8886585235595703,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.029827935621142387,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 2.8802051544189453,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.029231170192360878,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 2.876192092895508,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.029604196548461914,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 2.8420169353485107,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.030229991301894188,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 2.86871075630188,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.0357736200094223,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.852806568145752,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 4.496274948120117,
+ "eval_runtime": 58.0976,
+ "eval_samples_per_second": 42.033,
+ "eval_steps_per_second": 1.325,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.03180048242211342,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 2.85683012008667,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.030023347586393356,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 2.814507484436035,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.03346574679017067,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 2.8209877014160156,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.030521582812070847,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 2.7859511375427246,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.033295996487140656,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 2.7879018783569336,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.029781939461827278,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 2.801133155822754,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.033207762986421585,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 2.8046321868896484,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.03027796931564808,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 2.8235392570495605,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.03096267767250538,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 2.8181068897247314,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.030886022374033928,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 2.7954399585723877,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.030112111940979958,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 2.78225040435791,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.03067166730761528,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 2.8147826194763184,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.029821906238794327,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 2.811027765274048,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.030433299019932747,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 2.8178303241729736,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.02960469201207161,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 2.802370548248291,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.029496584087610245,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 2.8156142234802246,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.029699044302105904,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 2.7911806106567383,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.030225902795791626,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 2.7872819900512695,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.02939198724925518,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 2.800001621246338,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.03059281036257744,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 2.8127222061157227,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.02927992306649685,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 2.8320517539978027,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.03038649819791317,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 2.8169116973876953,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.029806654900312424,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 2.8279008865356445,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.029620900750160217,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 2.8212599754333496,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.029498178511857986,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 2.8112261295318604,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.02918926067650318,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 2.8290767669677734,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.029827842488884926,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 2.8396403789520264,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.029516329988837242,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 2.840583324432373,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.029057780280709267,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 2.835113525390625,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.029345065355300903,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 2.8197970390319824,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.02873479388654232,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 2.8052587509155273,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.029737083241343498,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 2.819146156311035,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.0292612686753273,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 2.8543701171875,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.028945421800017357,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 2.804084300994873,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.029137294739484787,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 2.7658708095550537,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.02911440096795559,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 2.8134429454803467,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.028932519257068634,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 2.810455322265625,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.029404157772660255,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 2.802886962890625,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.028970446437597275,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 2.8426849842071533,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.029187830165028572,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 2.8366644382476807,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.029293285682797432,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 2.8261566162109375,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.029366638511419296,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 2.8125827312469482,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.02900587022304535,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 2.839684009552002,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.029162462800741196,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 2.8331997394561768,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.029060618951916695,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 2.8627192974090576,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.02887798845767975,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 2.859687089920044,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.02894580364227295,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 2.8195912837982178,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.029554955661296844,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 2.852871894836426,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.02822914719581604,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 2.8671388626098633,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.02954651229083538,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 2.8524365425109863,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.02902315929532051,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 2.843954563140869,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.0290356632322073,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 2.8504233360290527,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.029262971132993698,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 2.8657913208007812,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.029469065368175507,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 2.82297945022583,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.029609832912683487,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 2.856297016143799,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.029690174385905266,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 2.8302533626556396,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.03018389455974102,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 2.8542540073394775,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.029053693637251854,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 2.8432576656341553,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.028549278154969215,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 2.8324697017669678,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.030014613643288612,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 2.800536870956421,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.02949996292591095,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 2.835141897201538,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.028825247660279274,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 2.8263447284698486,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.029436098411679268,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 2.8522257804870605,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.029510991647839546,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 2.851526975631714,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.028975380584597588,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 2.820462465286255,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.029128702357411385,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 2.861868381500244,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.02951551228761673,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 2.875210762023926,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.028196118772029877,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 2.8345048427581787,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.029215430840849876,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 2.84328556060791,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.02857460267841816,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 2.8448305130004883,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.02842608466744423,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 2.860792636871338,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03630338981747627,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.8723089694976807,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 4.522201061248779,
+ "eval_runtime": 57.3764,
+ "eval_samples_per_second": 42.561,
+ "eval_steps_per_second": 1.342,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.030005959793925285,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 2.7629683017730713,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.028917059302330017,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 2.7894129753112793,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.029286719858646393,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 2.8119699954986572,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.02967856638133526,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 2.8115768432617188,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.029460743069648743,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 2.7809553146362305,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.028186975046992302,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 2.818805456161499,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.02975514344871044,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 2.816215991973877,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.029532121494412422,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 2.782461166381836,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.02879161946475506,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 2.768501043319702,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.029776159673929214,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 2.7995424270629883,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.028667304664850235,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 2.773409843444824,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.028793245553970337,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 2.7938318252563477,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.029090600088238716,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 2.823000907897949,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.028883792459964752,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 2.8175344467163086,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.028518877923488617,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 2.792815685272217,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.029769321903586388,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 2.831251382827759,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.029009172692894936,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 2.8345084190368652,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.028994156047701836,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 2.7925689220428467,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.029505716636776924,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 2.814758777618408,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.028683487325906754,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 2.7954044342041016,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.029633600264787674,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 2.794569969177246,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.028943300247192383,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 2.803800106048584,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.029026521369814873,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 2.760383129119873,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.029057513922452927,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 2.8081226348876953,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.02906658686697483,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 2.8183770179748535,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.02926407754421234,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 2.831427574157715,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.029171453788876534,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 2.826587200164795,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.029656723141670227,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 2.7633590698242188,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.029490994289517403,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 2.7932424545288086,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.0291811041533947,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 2.8346633911132812,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.029164446517825127,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 2.8193140029907227,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.029394259676337242,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 2.8272342681884766,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.029221080243587494,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 2.79146146774292,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.028165455907583237,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 2.7826766967773438,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.0296048391610384,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 2.8119699954986572,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.028316112235188484,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 2.824860095977783,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02878737449645996,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 2.811997890472412,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029061628505587578,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 2.8370048999786377,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028607595711946487,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 2.8093791007995605,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.028144560754299164,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 2.8236052989959717,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.028813332319259644,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 2.838690996170044,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.028437303379178047,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 2.800854206085205,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.028567243367433548,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 2.8301069736480713,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.029187796637415886,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 2.823208808898926,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.028203707188367844,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 2.804643154144287,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.02822236344218254,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 2.8132784366607666,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.028921568766236305,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 2.8215417861938477,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.02867986261844635,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 2.8318984508514404,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.028849760070443153,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 2.8209228515625,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.028631621971726418,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 2.8219287395477295,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.028188595548272133,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 2.814027786254883,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.028504883870482445,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 2.847404956817627,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.02826615609228611,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 2.808562755584717,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02844209223985672,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 2.827244520187378,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.028765909373760223,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 2.81117844581604,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.028130069375038147,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 2.852351188659668,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.028408419340848923,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 2.847090244293213,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.028523078188300133,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 2.821716785430908,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.028461646288633347,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 2.816019296646118,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.028531840071082115,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 2.8226165771484375,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.028871113434433937,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 2.831231117248535,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.0285959430038929,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 2.8257393836975098,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02841980569064617,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 2.8298444747924805,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.028909973800182343,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 2.808438777923584,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.02884981967508793,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 2.815157890319824,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.02836848609149456,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 2.792750835418701,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.028464308008551598,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 2.7594008445739746,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.02909877710044384,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 2.844888925552368,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.0285524632781744,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 2.8162052631378174,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.02916364185512066,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 2.8175275325775146,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.028360974043607712,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 2.8322806358337402,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.0366419292986393,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.8780605792999268,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 4.539337635040283,
+ "eval_runtime": 57.482,
+ "eval_samples_per_second": 42.483,
+ "eval_steps_per_second": 1.34,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030007537454366684,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 2.7774059772491455,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.028395716100931168,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 2.8240408897399902,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.029259607195854187,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 2.7780704498291016,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.029434163123369217,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 2.802172899246216,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.02914051152765751,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 2.746731758117676,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.02971787191927433,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 2.7998170852661133,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.029050175100564957,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 2.786863327026367,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.030421903356909752,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 2.789545774459839,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.029742848128080368,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 2.827561855316162,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.03001682087779045,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 2.8226547241210938,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.030296841636300087,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 2.795665740966797,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.02908152900636196,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 2.7598729133605957,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.029059268534183502,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 2.786234140396118,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.02878507599234581,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 2.7754065990448,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.02881552092730999,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 2.779616355895996,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.02847588248550892,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 2.7556540966033936,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.028446536511182785,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 2.7858777046203613,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.028488771989941597,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 2.8308897018432617,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.028518928214907646,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 2.798438310623169,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.02807544358074665,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 2.7854204177856445,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.028449799865484238,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 2.787806510925293,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.02806367352604866,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 2.773944139480591,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.028127942234277725,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 2.795468807220459,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.028206733986735344,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 2.794552803039551,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.02769204042851925,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 2.78456974029541,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.028235256671905518,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 2.8258988857269287,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.028410933911800385,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 2.8145394325256348,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.028073569759726524,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 2.739804744720459,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.027985481545329094,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 2.7385730743408203,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.0285568255931139,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 2.790156364440918,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.028381718322634697,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 2.7715389728546143,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028455479070544243,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 2.804957628250122,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.028330320492386818,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 2.813432216644287,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.02832239493727684,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 2.8087422847747803,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.028423218056559563,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 2.790351390838623,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.028114303946495056,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 2.8367419242858887,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.028525900095701218,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 2.8024988174438477,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.027810173109173775,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 2.8156440258026123,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.028172023594379425,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 2.801828384399414,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.027999311685562134,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 2.7917897701263428,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.02863648533821106,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 2.7998974323272705,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.027806704863905907,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 2.7955031394958496,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.028083806857466698,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 2.8160462379455566,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02817711979150772,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 2.7673726081848145,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.027995774522423744,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 2.787623882293701,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.028293436393141747,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 2.848489284515381,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.028710173442959785,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 2.8186140060424805,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.028271062299609184,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 2.822369337081909,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.02830388769507408,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 2.8050756454467773,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.02831537462770939,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 2.8161096572875977,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.028237098827958107,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 2.7897520065307617,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.027944844216108322,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 2.816229820251465,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.028575003147125244,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 2.8069968223571777,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.028569547459483147,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 2.78981614112854,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.028383128345012665,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 2.820969343185425,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.027872422710061073,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 2.781301498413086,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.028342951089143753,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 2.816469192504883,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.028373923152685165,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 2.799290657043457,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.02807556465268135,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 2.814924716949463,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.02789521962404251,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 2.7950439453125,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.028297580778598785,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 2.8535611629486084,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.02745051123201847,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 2.7812037467956543,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.027246123179793358,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 2.778583288192749,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.027801746502518654,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 2.812328815460205,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.027761677280068398,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 2.829758882522583,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.028340548276901245,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 2.8203043937683105,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.027355236932635307,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 2.777925968170166,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027668653056025505,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 2.789498805999756,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02823868952691555,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 2.7765135765075684,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.02812831476330757,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 2.774756669998169,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.027789659798145294,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 2.8317346572875977,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.035616710782051086,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.8123791217803955,
+ "step": 3240
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.3813260125995008e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-3240/training_args.bin b/runs/l2r90-baseline/checkpoint-3240/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3240/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-360/chat_template.jinja b/runs/l2r90-baseline/checkpoint-360/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-360/config.json b/runs/l2r90-baseline/checkpoint-360/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-360/generation_config.json b/runs/l2r90-baseline/checkpoint-360/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-360/model.safetensors b/runs/l2r90-baseline/checkpoint-360/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b2093bf098854f2aa2361e88b041436e51ed2c95
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5518c6563c8ab5204126d1b92837f3a9273693a6812f1319d2fa9d9baff16348
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-360/optimizer.pt b/runs/l2r90-baseline/checkpoint-360/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5741cbf54e5f61b8adb0ae9c46f767851689c892
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fac2b4effb5f939d5027f6ada65805f1279975071f7d418ee92347c24b3fa06d
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c3dcf55387a71ee2a6afda87f94800657089f6cc
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e5cf3bac380addaa516f941e6a276bb95c92ecfb3c08371ff9fba4b2faac6df0
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..26b693a57bc20ed89f333203ee3df5b86d2cc75b
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:219db464e75bbf2b6c41a71470b6182f4de4159f65fd71e3323fd50c658eea3a
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..9f43ebb920eaa7f2483f205604373ffec78981af
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:03b8316374c36bf601b0d74a53807a0d460d535e42771ed5f0f6db088f66ab63
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..c89f029f8f7c32ec2bc7c6471885e82ebfe5322e
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6d6c7bdfadb70cba54ce8abee8b9fbcd1d5870419c1ccbd2e7362a47c09ba74e
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-360/scheduler.pt b/runs/l2r90-baseline/checkpoint-360/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d63b22e047ab13a6c29adc08f78969e1d6e569ff
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2802ceffe363a15cca2d2192020ec6c680a5afa8cc5d08bb488000d79d1a8eb0
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-360/tokenizer.json b/runs/l2r90-baseline/checkpoint-360/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-360/trainer_state.json b/runs/l2r90-baseline/checkpoint-360/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..8632b646315bb3cfa011301c5d864aba8e47cffb
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/trainer_state.json
@@ -0,0 +1,2586 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 360,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.534806680666112e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-360/training_args.bin b/runs/l2r90-baseline/checkpoint-360/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-360/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja b/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-3600/config.json b/runs/l2r90-baseline/checkpoint-3600/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-3600/generation_config.json b/runs/l2r90-baseline/checkpoint-3600/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-3600/model.safetensors b/runs/l2r90-baseline/checkpoint-3600/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8744a97f301b6417af7c4e09ffb07028f35e2f44
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4b0079515fa8e03d4ede34fbbc1c4a85937b053aa209c5fceaf8a4b94716babe
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-3600/optimizer.pt b/runs/l2r90-baseline/checkpoint-3600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..50dc69f9064de595987eb41f1c4d7c2b98002fb1
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e8f6d37ab699c0adaaf7cf682e015df8719cbdc2d1af74aab7e49969edb61382
+size 1166827898
diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1a514451b6a50ad9b6de0b7fc4a965cb83108e58
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a6ec8d61675b917843283ea16decabe601854cecb62727b8d6fd0736b1d8e982
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4b5687b836d8e7154f2f48e4093e48f7463e53d0
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bf4a2454d9040e2fdfe5d3ef7263fb3bdbcaa970303baa7413c80aed2f694675
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..31f9a655287c5cfc3f33b435d1688e36b4b32220
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e25cf1e6918353e4d123a8988a16033aeb3a27afe509656b117bb952c8698b33
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..276e9dddc44b2549c7e379c8e61b12187f2dd7e3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2366911094763ba47ae31b700e5549017c3b44f5913088fc81788f708e2fa6b0
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-3600/scheduler.pt b/runs/l2r90-baseline/checkpoint-3600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..dff04aa3b724ae7dec72031a9c3d0ecf0b695f1b
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:187927afe10d5874e052562039160b9509612fe972489c80a97620ccae701300
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-3600/tokenizer.json b/runs/l2r90-baseline/checkpoint-3600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-3600/trainer_state.json b/runs/l2r90-baseline/checkpoint-3600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..3af15bc60ef2d14f943069807fe41aff3afe2684
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/trainer_state.json
@@ -0,0 +1,25618 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 50.0,
+ "eval_steps": 500,
+ "global_step": 3600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 4.4757890701293945,
+ "eval_runtime": 57.7451,
+ "eval_samples_per_second": 42.289,
+ "eval_steps_per_second": 1.333,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.029025450348854065,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 4.288647174835205,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.029403962194919586,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 4.288555145263672,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.03265019878745079,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 4.292625904083252,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.0387202650308609,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 4.238926887512207,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.04554206505417824,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 4.29966402053833,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.036231476813554764,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 4.304398536682129,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.038703009486198425,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 4.2945356369018555,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.03224619850516319,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 4.261351585388184,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.029722856357693672,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 4.303143501281738,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.02764045260846615,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 4.243679523468018,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.0312200877815485,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 4.279904365539551,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.027349065989255905,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 4.270228385925293,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.030385233461856842,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 4.301590442657471,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.03292274847626686,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 4.301828384399414,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.03265977278351784,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 4.286200523376465,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.03177836537361145,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 4.296319007873535,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.028879934921860695,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 4.261706829071045,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.02855812944471836,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 4.273903846740723,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.03037901036441326,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 4.248080253601074,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.029279105365276337,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 4.281311988830566,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.02592580020427704,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 4.204848289489746,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.021899603307247162,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 4.232453346252441,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.019704442471265793,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 4.253462314605713,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.019550519064068794,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 4.307737350463867,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.020320113748311996,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 4.257411956787109,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.02118559181690216,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 4.262089729309082,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.02050616219639778,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 4.2490644454956055,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019793255254626274,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 4.267400741577148,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.020760254934430122,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 4.2532429695129395,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.019524799659848213,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 4.267422199249268,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.020166637375950813,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 4.214296817779541,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.01964394748210907,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 4.252087116241455,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.01690433733165264,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 4.232336044311523,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.017682762816548347,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 4.213962554931641,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.01688133366405964,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 4.199461460113525,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.019166061654686928,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 4.228936672210693,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.024065518751740456,
+ "learning_rate": 0.000549983495527656,
+ "loss": 4.214227676391602,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.02518288791179657,
+ "learning_rate": 0.000549835338229793,
+ "loss": 4.256412029266357,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.026093967258930206,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 4.216564178466797,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.023775635287165642,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 4.226430892944336,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.018810907378792763,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 4.190013885498047,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.018974872305989265,
+ "learning_rate": 0.000549240721526684,
+ "loss": 4.246231555938721,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.02069833129644394,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 4.198457717895508,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.02097306028008461,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 4.212507247924805,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022165542468428612,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 4.2043843269348145,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.019068919122219086,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 4.197460651397705,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01701812446117401,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 4.220056533813477,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.01629726029932499,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 4.214868545532227,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018661806359887123,
+ "learning_rate": 0.000548192508857563,
+ "loss": 4.227047920227051,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.02005891129374504,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 4.197265625,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.02591128647327423,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 4.211041450500488,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.032659754157066345,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 4.230195045471191,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.04057808965444565,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 4.223843097686768,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.04103454574942589,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 4.197145462036133,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.03810729831457138,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 4.237957000732422,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.028838301077485085,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 4.206722259521484,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.02436191774904728,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 4.1983137130737305,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.022839387878775597,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 4.20530366897583,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.02375260926783085,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 4.213202476501465,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.02177741937339306,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 4.200539588928223,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.02351468615233898,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 4.198091506958008,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.023375794291496277,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 4.193244934082031,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.02479805052280426,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 4.2250871658325195,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.026894841343164444,
+ "learning_rate": 0.000545913797265339,
+ "loss": 4.241791248321533,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.030582403764128685,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 4.203666687011719,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.032222259789705276,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 4.209629535675049,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.03181060031056404,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 4.200459957122803,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02687576413154602,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 4.211819648742676,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.025988668203353882,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 4.198040962219238,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.02937682718038559,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 4.222935676574707,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.03043951839208603,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 4.2031755447387695,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.030401621013879776,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 4.198380470275879,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 4.368444919586182,
+ "eval_runtime": 58.3009,
+ "eval_samples_per_second": 41.886,
+ "eval_steps_per_second": 1.321,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.03461585193872452,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 4.147369384765625,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.039084456861019135,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 4.124129295349121,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.04021848738193512,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 4.127362251281738,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.041000768542289734,
+ "learning_rate": 0.00054405910314802,
+ "loss": 4.1432061195373535,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.04008613899350166,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 4.169931888580322,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.039394307881593704,
+ "learning_rate": 0.000543747262591706,
+ "loss": 4.155813217163086,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.04374868795275688,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 4.159891128540039,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.0443582646548748,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 4.157362937927246,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.04601646214723587,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 4.152097225189209,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.04327242821455002,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 4.170074462890625,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.0413462370634079,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 4.156710147857666,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.036855220794677734,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 4.138053894042969,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.03781217709183693,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 4.172306060791016,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.03886301815509796,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 4.135544776916504,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.03730994835495949,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 4.171654224395752,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.03271950036287308,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 4.169906139373779,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.03292673081159592,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 4.163660049438477,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.029366906732320786,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 4.161055088043213,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.023805437609553337,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 4.172390937805176,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.023178137838840485,
+ "learning_rate": 0.000541542711134479,
+ "loss": 4.146788597106934,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.02403450198471546,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 4.124364852905273,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.025009717792272568,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 4.12275505065918,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.023431289941072464,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 4.134739398956299,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.023803815245628357,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 4.130463123321533,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.021219994872808456,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 4.124577522277832,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.01899666152894497,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 4.167086124420166,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.01889493502676487,
+ "learning_rate": 0.000540426287667833,
+ "loss": 4.152744293212891,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.019546914845705032,
+ "learning_rate": 0.000540266032410741,
+ "loss": 4.122655868530273,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.02085448056459427,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 4.141949653625488,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.019959403201937675,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 4.131907939910889,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.01885392889380455,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 4.118953227996826,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.018216131255030632,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 4.096459865570068,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.019385207444429398,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 4.111050128936768,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.018676603212952614,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 4.156215667724609,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.01876339502632618,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 4.1107282638549805,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.01831623539328575,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 4.154116630554199,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.018896661698818207,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 4.093320846557617,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.019582601264119148,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 4.106500625610352,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.020313601940870285,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 4.071961879730225,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.020385954529047012,
+ "learning_rate": 0.000538328095851792,
+ "loss": 4.129006862640381,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.02231123298406601,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 4.067756652832031,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.022957179695367813,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 4.088230133056641,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.021438462659716606,
+ "learning_rate": 0.000537839336909727,
+ "loss": 4.128036022186279,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.01893841102719307,
+ "learning_rate": 0.00053767603867558,
+ "loss": 4.155407905578613,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.018539676442742348,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 4.101880073547363,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.01881185919046402,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 4.118906021118164,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.021657396107912064,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 4.091935157775879,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.023293187841773033,
+ "learning_rate": 0.000537020956314485,
+ "loss": 4.105803489685059,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.021306317299604416,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 4.0610809326171875,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.022078553214669228,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 4.09867525100708,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.02266664244234562,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 4.1027326583862305,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.02091757394373417,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 4.080266952514648,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.020374659448862076,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 4.088479042053223,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.02294750325381756,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 4.084364891052246,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.02337353676557541,
+ "learning_rate": 0.000535867308253152,
+ "loss": 4.125696182250977,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.022643696516752243,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 4.097160339355469,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.021080605685710907,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 4.102363109588623,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.02269427664577961,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 4.089844703674316,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.02195413038134575,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 4.132293701171875,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.01805616170167923,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 4.0842814445495605,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.018048783764243126,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 4.109591007232666,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.0199508648365736,
+ "learning_rate": 0.000534704466989222,
+ "loss": 4.0908074378967285,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.020186785608530045,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 4.077959060668945,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.022526193410158157,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 4.109979629516602,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.025224441662430763,
+ "learning_rate": 0.000534203303565406,
+ "loss": 4.047533988952637,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.027292724698781967,
+ "learning_rate": 0.000534035876358542,
+ "loss": 4.09566593170166,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.025669090449810028,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 4.0606160163879395,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.02791150100529194,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 4.097081184387207,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.03294133022427559,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 4.121442794799805,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.027322066947817802,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 4.1188740730285645,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.024866372346878052,
+ "learning_rate": 0.000533195950905842,
+ "loss": 4.082042694091797,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.02901027724146843,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 4.110199451446533,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 4.283970355987549,
+ "eval_runtime": 58.4003,
+ "eval_samples_per_second": 41.815,
+ "eval_steps_per_second": 1.318,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.028958842158317566,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 4.040733337402344,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.03491180017590523,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 4.0473432540893555,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.04025818780064583,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 4.021976470947266,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.03998899459838867,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 4.04901123046875,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.042165111750364304,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 4.050230979919434,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.03511597216129303,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 4.0989990234375,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.04064979776740074,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 4.051448822021484,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.047536592930555344,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 4.064278602600098,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.05592669919133186,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 4.045516014099121,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.04262633994221687,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 4.060666084289551,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.03997781500220299,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 4.065315246582031,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.038152433931827545,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 4.069721221923828,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.0367623008787632,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 4.080986976623535,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.0358598455786705,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 4.065290927886963,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.039188165217638016,
+ "learning_rate": 0.000530477115403131,
+ "loss": 4.024333477020264,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.036839909851551056,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 4.067532539367676,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.03456985577940941,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 4.048443794250488,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.039487943053245544,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 4.0585737228393555,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.03415827825665474,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 4.06104850769043,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.03209620714187622,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 4.009165287017822,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.028947360813617706,
+ "learning_rate": 0.0005294454257704,
+ "loss": 4.067384243011475,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.026836825534701347,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 4.046419620513916,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.024095216765999794,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 4.019008636474609,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.020965851843357086,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 4.020613670349121,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.02250855043530464,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 4.031634330749512,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.021385207772254944,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 4.086246013641357,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.021104589104652405,
+ "learning_rate": 0.000528407168205483,
+ "loss": 4.033970355987549,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.019584037363529205,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 4.025118827819824,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.02107258513569832,
+ "learning_rate": 0.000528059628009464,
+ "loss": 4.020823955535889,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.020438389852643013,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 4.016323566436768,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.01881488971412182,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 3.9991841316223145,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.01763618178665638,
+ "learning_rate": 0.000527536958117778,
+ "loss": 4.0062150955200195,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.017308862879872322,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 3.946723461151123,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.01706845872104168,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 4.056562900543213,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.018664294853806496,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 4.024652481079102,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.018523404374718666,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 3.962064743041992,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.017408935353159904,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 4.008697509765625,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.01588570512831211,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 4.011664390563965,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.015084310434758663,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 3.996224880218506,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.01637054979801178,
+ "learning_rate": 0.000526135222238964,
+ "loss": 4.015972137451172,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.01673802360892296,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 3.975414276123047,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.016544019803404808,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 4.012955188751221,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.017455460503697395,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 4.010180473327637,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.017625460401177406,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 4.001679420471191,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.01732795126736164,
+ "learning_rate": 0.000525253290006091,
+ "loss": 4.006107330322266,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.016654646024107933,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 4.045180797576904,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.015085658989846706,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 3.994246482849121,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.015112568624317646,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 4.004963397979736,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.014882124960422516,
+ "learning_rate": 0.000524544520190982,
+ "loss": 4.029565811157227,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.016012471169233322,
+ "learning_rate": 0.000524366881143897,
+ "loss": 4.018208026885986,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.015761924907565117,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 3.9911205768585205,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.017770355567336082,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 4.028733253479004,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.01959751546382904,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 4.003322124481201,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.018713824450969696,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 3.991718292236328,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.01992208883166313,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 4.002507209777832,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.021671291440725327,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 4.021344184875488,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024282798171043396,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 3.9899158477783203,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.023294344544410706,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 4.046627044677734,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.02029707282781601,
+ "learning_rate": 0.000522760121133566,
+ "loss": 4.022454261779785,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.02151239849627018,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 4.005476951599121,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.02320881187915802,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 4.012726783752441,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.026484191417694092,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 3.9904391765594482,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.026279661804437637,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 3.9738848209381104,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.024299729615449905,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 4.003316402435303,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.02337595447897911,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 3.954984188079834,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.02416428178548813,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 4.0317063331604,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.022488107904791832,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 3.9736647605895996,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.019700225442647934,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 3.979867458343506,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.020382333546876907,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 3.995251178741455,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.020149298012256622,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 3.9987268447875977,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.019790509715676308,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 3.9838991165161133,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.022096451371908188,
+ "learning_rate": 0.000520413954218197,
+ "loss": 3.995800256729126,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 4.214347839355469,
+ "eval_runtime": 58.8872,
+ "eval_samples_per_second": 41.469,
+ "eval_steps_per_second": 1.308,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.02203402668237686,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 3.915332078933716,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.02336675301194191,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 3.914801597595215,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.027339637279510498,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 3.9277780055999756,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.026872653514146805,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 3.9160094261169434,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.026558255776762962,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 3.9084300994873047,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.02644367143511772,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 3.9328718185424805,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.024677276611328125,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 3.88993501663208,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.023194244131445885,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 3.923567056655884,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.024653147906064987,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 3.8740999698638916,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.026041202247142792,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 3.9542148113250732,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.029586758464574814,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 3.9241347312927246,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.0325787179172039,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 3.9656593799591064,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.03183380514383316,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 3.950648307800293,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.03434471786022186,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 3.9590439796447754,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.030099432915449142,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 3.9518113136291504,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.025950564071536064,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 3.8908939361572266,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.029959935694932938,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 3.945037364959717,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.029551763087511063,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 3.923959732055664,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.03538523241877556,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 3.9395527839660645,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.037924155592918396,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 3.945115089416504,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.036773014813661575,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 3.9484646320343018,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.04125838354229927,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 4.000699043273926,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.044125135987997055,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 3.9828028678894043,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.040853384882211685,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 3.957505702972412,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.03501008450984955,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 3.9360849857330322,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.04082556813955307,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 3.95941424369812,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.045432962477207184,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 3.9666922092437744,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.04048244655132294,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 3.984555244445801,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.031016314402222633,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 3.945204257965088,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.03210771828889847,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 3.976186752319336,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.03130023553967476,
+ "learning_rate": 0.000514700389506765,
+ "loss": 3.915727138519287,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.031193193048238754,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 3.961866855621338,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.027518954128026962,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 3.9742956161499023,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.02751155197620392,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 3.949662208557129,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.026033353060483932,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 3.9327926635742188,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.02579621411859989,
+ "learning_rate": 0.000513763379483416,
+ "loss": 3.9334418773651123,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.023009559139609337,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 3.969707489013672,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.02196447364985943,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 3.9084434509277344,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.02206484042108059,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 3.922117233276367,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.023360205814242363,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 3.965158224105835,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.02162669226527214,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 3.9519729614257812,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.019719846546649933,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 3.9195961952209473,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.018988260999321938,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 3.937926769256592,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.019278401508927345,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 3.9126808643341064,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.019101005047559738,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 3.8993749618530273,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.019156700000166893,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 3.9531068801879883,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.019944407045841217,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 3.9645180702209473,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.01826174184679985,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 3.948477268218994,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.0191368218511343,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 3.944648027420044,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.017768343910574913,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 3.9584474563598633,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.01955590397119522,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 3.8936800956726074,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.02100536972284317,
+ "learning_rate": 0.000510736513223685,
+ "loss": 3.93919038772583,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.020873935893177986,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 3.929244041442871,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.021876046434044838,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 3.9264731407165527,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.021726684644818306,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 3.923135280609131,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.019308188930153847,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 3.9062135219573975,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01917017623782158,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 3.898810386657715,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017933931201696396,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 3.8899734020233154,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.015131795778870583,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 3.9452900886535645,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.017190825194120407,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 3.9124746322631836,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.01618487387895584,
+ "learning_rate": 0.000509015031826403,
+ "loss": 3.91654109954834,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.014920198358595371,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 3.9144535064697266,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01615205965936184,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 3.917003631591797,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.015370192006230354,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 3.8910603523254395,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.015290996059775352,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 3.9122695922851562,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.015913106501102448,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 3.9304556846618652,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.01639600843191147,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 3.8974502086639404,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.017010601237416267,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 3.9723148345947266,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.017778996378183365,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 3.8905391693115234,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.020029611885547638,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 3.934976100921631,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.01931857503950596,
+ "learning_rate": 0.000507086511366679,
+ "loss": 3.9152774810791016,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.019750485196709633,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 3.877650499343872,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 4.161914825439453,
+ "eval_runtime": 58.5217,
+ "eval_samples_per_second": 41.728,
+ "eval_steps_per_second": 1.316,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.02073049545288086,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 3.8331246376037598,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.02272389829158783,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 3.8847155570983887,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.02139078453183174,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 3.8276169300079346,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.02043229155242443,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 3.8804001808166504,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.02345564216375351,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 3.851609468460083,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.023448185995221138,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 3.8446507453918457,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.024826066568493843,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 3.86712384223938,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.02491980977356434,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 3.8374528884887695,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.023002834990620613,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 3.783214569091797,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.02123952843248844,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 3.8463428020477295,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.01981225423514843,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 3.8307082653045654,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.01915367692708969,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 3.8190226554870605,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.01890859752893448,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 3.8428220748901367,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.019150376319885254,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 3.830247402191162,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.020099453628063202,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 3.838806629180908,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.021136781200766563,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 3.8344273567199707,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.022857150062918663,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 3.8317880630493164,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.024891739711165428,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 3.837855577468872,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.02595464698970318,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 3.846050262451172,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.02722409926354885,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 3.8164305686950684,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.024486860260367393,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 3.8099875450134277,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.021551169455051422,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 3.825925827026367,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.02121506817638874,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 3.870727777481079,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.01948230341076851,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 3.8035314083099365,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.01896044984459877,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 3.84665584564209,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.020620400086045265,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 3.8338310718536377,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.020855262875556946,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 3.8781003952026367,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.02153870463371277,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 3.8783490657806396,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.023751024156808853,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 3.846198558807373,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.023987066000699997,
+ "learning_rate": 0.000501004302813408,
+ "loss": 3.834296226501465,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.023626182228326797,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 3.8591227531433105,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.025707192718982697,
+ "learning_rate": 0.000500606587074199,
+ "loss": 3.8766489028930664,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.027723029255867004,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 3.88264799118042,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.0278895553201437,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 3.8932995796203613,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.025622662156820297,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 3.8734612464904785,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.02639615908265114,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 3.786813735961914,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.026765599846839905,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 3.853966474533081,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.02823319099843502,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 3.8321340084075928,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.023394955322146416,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 3.8527331352233887,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.023874662816524506,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 3.8822879791259766,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.02301311306655407,
+ "learning_rate": 0.000498809004003543,
+ "loss": 3.846186876296997,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.022630592808127403,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 3.867992639541626,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.02424517087638378,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 3.867293357849121,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.024116624146699905,
+ "learning_rate": 0.000498206964168724,
+ "loss": 3.846282958984375,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.02187768742442131,
+ "learning_rate": 0.000498005969458628,
+ "loss": 3.8429484367370605,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.01879809983074665,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 3.863215446472168,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.020426234230399132,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 3.825230121612549,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01992812193930149,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 3.851665496826172,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.020693328231573105,
+ "learning_rate": 0.000497200421111647,
+ "loss": 3.906999111175537,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.021137919276952744,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 3.8369035720825195,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.023285407572984695,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 3.8658390045166016,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.0219135619699955,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 3.873072862625122,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.020101983100175858,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 3.847805976867676,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.01875176839530468,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 3.7976622581481934,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.01913360133767128,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 3.853851318359375,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.016863619908690453,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 3.877378225326538,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.017802366986870766,
+ "learning_rate": 0.000495581824580724,
+ "loss": 3.861356019973755,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.01981814205646515,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 3.88613224029541,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.017508897930383682,
+ "learning_rate": 0.000495175620586125,
+ "loss": 3.8202030658721924,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.019240548834204674,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 3.853827714920044,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.022305598482489586,
+ "learning_rate": 0.000494768797261945,
+ "loss": 3.885148763656616,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.02531025931239128,
+ "learning_rate": 0.000494565153760897,
+ "loss": 3.8412041664123535,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.024474849924445152,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 3.865644931793213,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.020430924370884895,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 3.8649442195892334,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.01906431093811989,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 3.8462204933166504,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.021118057891726494,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 3.8652522563934326,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.020337693393230438,
+ "learning_rate": 0.00049354462443268,
+ "loss": 3.8506221771240234,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.019090797752141953,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 3.881237506866455,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.020973797887563705,
+ "learning_rate": 0.000493135336920878,
+ "loss": 3.821274757385254,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.021984362974762917,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 3.8793835639953613,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.0215684212744236,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 3.8259663581848145,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.021673865616321564,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 3.8505711555480957,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.123871326446533,
+ "eval_runtime": 58.8357,
+ "eval_samples_per_second": 41.505,
+ "eval_steps_per_second": 1.309,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.0203359667211771,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 3.785353660583496,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.023197868838906288,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 3.7479515075683594,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.02314298041164875,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 3.7697315216064453,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.0256908368319273,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 3.79056978225708,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.026193121448159218,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 3.7722532749176025,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.027120303362607956,
+ "learning_rate": 0.000491285979372924,
+ "loss": 3.7991981506347656,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.0266975536942482,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 3.778301477432251,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.02806396596133709,
+ "learning_rate": 0.000490873338088198,
+ "loss": 3.7578704357147217,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.02893839031457901,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 3.7401275634765625,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.031170370057225227,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 3.780686616897583,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.03364266827702522,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 3.794442653656006,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.030121006071567535,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 3.779569149017334,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.03296204283833504,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 3.7960336208343506,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.03381749242544174,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 3.7959280014038086,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.035383883863687515,
+ "learning_rate": 0.000489424334303338,
+ "loss": 3.8008625507354736,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.03759761154651642,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 3.7997028827667236,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.030651796609163284,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 3.81191349029541,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.0262477844953537,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 3.8046622276306152,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.030398324131965637,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 3.7903637886047363,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.030515259131789207,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 3.8111231327056885,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.031686048954725266,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 3.781445026397705,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.025968654081225395,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 3.788196563720703,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.025340190157294273,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 3.763331413269043,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.024738596752285957,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 3.8005969524383545,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.021990075707435608,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 3.795844078063965,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.02191784791648388,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 3.8062384128570557,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.023678259924054146,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 3.7760679721832275,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.024137988686561584,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 3.7917492389678955,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.02673223614692688,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 3.8006649017333984,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.02395169623196125,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 3.7902212142944336,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.02410844899713993,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 3.8073291778564453,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.028109831735491753,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 3.80116605758667,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.025538964197039604,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 3.791151523590088,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.02393815852701664,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 3.7766106128692627,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.024659741669893265,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 3.789135217666626,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.02668776921927929,
+ "learning_rate": 0.000485033394934934,
+ "loss": 3.797905445098877,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.0258641317486763,
+ "learning_rate": 0.000484822676912974,
+ "loss": 3.783323287963867,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.024846022948622704,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 3.7976717948913574,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.024981874972581863,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 3.818732738494873,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.02556881681084633,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 3.7821896076202393,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.02558571845293045,
+ "learning_rate": 0.000483978341612154,
+ "loss": 3.8143680095672607,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.02361566200852394,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 3.7710189819335938,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.022570716217160225,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 3.775564670562744,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.022478004917502403,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 3.7920689582824707,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.021241728216409683,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 3.7857837677001953,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.018921511247754097,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 3.7659268379211426,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01926037296652794,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 3.7527129650115967,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.018743667751550674,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 3.767991065979004,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.017273949459195137,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 3.7834174633026123,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.017054244875907898,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 3.7775752544403076,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.017190931364893913,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 3.7709648609161377,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.01665034517645836,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 3.7575321197509766,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.016932843253016472,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 3.7922799587249756,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01702750287950039,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 3.7729780673980713,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.01766599342226982,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 3.7876572608947754,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01785697042942047,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 3.8131699562072754,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01830688677728176,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 3.790778636932373,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.017348438501358032,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 3.7984330654144287,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.017999377101659775,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 3.7756147384643555,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.01801210641860962,
+ "learning_rate": 0.000479936075927473,
+ "loss": 3.789759874343872,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.018445398658514023,
+ "learning_rate": 0.000479721889242305,
+ "loss": 3.7683000564575195,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.017544370144605637,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 3.7776038646698,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.019834233447909355,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 3.7559571266174316,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01791626401245594,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 3.7746729850769043,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.017638562247157097,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 3.7933082580566406,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.018974050879478455,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 3.795713424682617,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.018694303929805756,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 3.7675955295562744,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.01788092777132988,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 3.786083698272705,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.01902497000992298,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 3.7976233959198,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.020558880642056465,
+ "learning_rate": 0.000477787822648007,
+ "loss": 3.7500152587890625,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.020093899220228195,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 3.703448534011841,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.019833799451589584,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 3.756059408187866,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.093987941741943,
+ "eval_runtime": 57.6894,
+ "eval_samples_per_second": 42.33,
+ "eval_steps_per_second": 1.335,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.022167526185512543,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 3.696444511413574,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.027874166145920753,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 3.713630199432373,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.030872784554958344,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 3.678295612335205,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.03246859833598137,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 3.7186388969421387,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.03667069226503372,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 3.7457995414733887,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.036286015063524246,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 3.6881775856018066,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.03413614630699158,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 3.6999263763427734,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.03384552150964737,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 3.6959803104400635,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.03227181360125542,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 3.7182722091674805,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.036366235464811325,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 3.7116851806640625,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.03573326766490936,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 3.741865873336792,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.03482096269726753,
+ "learning_rate": 0.000474756648870413,
+ "loss": 3.7195520401000977,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.03232625126838684,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 3.7443251609802246,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.03421664610505104,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 3.6867361068725586,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.03585013002157211,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 3.7492403984069824,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.029985329136252403,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 3.705209493637085,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.02729589305818081,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 3.711106538772583,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.029642216861248016,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 3.7274727821350098,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.02816903032362461,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 3.7181742191314697,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.027143625542521477,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 3.719235897064209,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.025802744552493095,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 3.771820545196533,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.026424532756209373,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 3.7516651153564453,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.029677648097276688,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 3.7058472633361816,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.031548649072647095,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 3.756992816925049,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.027667395770549774,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 3.742711067199707,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.027639545500278473,
+ "learning_rate": 0.000471698352726896,
+ "loss": 3.728883743286133,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.02950332500040531,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 3.7160212993621826,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.026302607730031013,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 3.735713481903076,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.022483112290501595,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 3.702686309814453,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.021729422733187675,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 3.7195816040039062,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.021159254014492035,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 3.684321880340576,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.02233424223959446,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 3.7233057022094727,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.02142048068344593,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 3.7083680629730225,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.022331461310386658,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 3.7408461570739746,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.020629866048693657,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 3.759129047393799,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.02174685336649418,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 3.72351336479187,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.021538110449910164,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 3.716416597366333,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.020800625905394554,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 3.726954936981201,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.02039588987827301,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 3.7336390018463135,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.02107219770550728,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 3.722311496734619,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.02068956382572651,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 3.767139434814453,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.020190054550766945,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 3.6960511207580566,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.018814411014318466,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 3.70375919342041,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.017568519338965416,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 3.756627321243286,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.018188195303082466,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 3.6964683532714844,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.018925746902823448,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 3.731172561645508,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.01690756157040596,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 3.6955318450927734,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016300756484270096,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 3.762632369995117,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.016405068337917328,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 3.75130558013916,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.017687877640128136,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 3.687530755996704,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.017524344846606255,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 3.738790273666382,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.01714804768562317,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 3.69827938079834,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.018042324110865593,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 3.71600341796875,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.018648479133844376,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 3.742591619491577,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.019198937341570854,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 3.739820957183838,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.01796373911201954,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 3.719025135040283,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01836620457470417,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 3.741434097290039,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.02192627638578415,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 3.7553882598876953,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.02147185243666172,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 3.7705790996551514,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.01855960488319397,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 3.704925537109375,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.01925475336611271,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 3.711491346359253,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.019796503707766533,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 3.7241992950439453,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.018511971458792686,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 3.6785435676574707,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.018214019015431404,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 3.7299962043762207,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.018912190571427345,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 3.773698329925537,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.01939854398369789,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 3.682321071624756,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.017859861254692078,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 3.7254951000213623,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.019249707460403442,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 3.7263433933258057,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.019221967086195946,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 3.7229225635528564,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01871953159570694,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 3.7248311042785645,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01747182011604309,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 3.699913740158081,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.019630730152130127,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 3.787783145904541,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.072091102600098,
+ "eval_runtime": 63.1589,
+ "eval_samples_per_second": 38.664,
+ "eval_steps_per_second": 1.219,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.022918934002518654,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 3.666675567626953,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.0264838095754385,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 3.6371259689331055,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.028463594615459442,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 3.657130241394043,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.02900792844593525,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 3.6365628242492676,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.028593607246875763,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 3.647258996963501,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.02878790907561779,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 3.682035446166992,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.02980183996260166,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 3.623297691345215,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.03375725448131561,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 3.6240830421447754,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.035045325756073,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 3.6372249126434326,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.03519923985004425,
+ "learning_rate": 0.000459203665939781,
+ "loss": 3.6157641410827637,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.03468473628163338,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 3.6553335189819336,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.034817393869161606,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 3.653806686401367,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.03549059107899666,
+ "learning_rate": 0.000458522886595554,
+ "loss": 3.6730189323425293,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.028001410886645317,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 3.662490129470825,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.02532520331442356,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 3.629936456680298,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.028215918689966202,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 3.6507444381713867,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.026823483407497406,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 3.661811113357544,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.0281058382242918,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 3.6805872917175293,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.02548503689467907,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 3.6586413383483887,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.024795135483145714,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 3.654989004135132,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.025501688942313194,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 3.6915903091430664,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.02377370186150074,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 3.660588264465332,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.022150425240397453,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 3.64870023727417,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.02415974996984005,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 3.6498191356658936,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.02140078693628311,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 3.703489065170288,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.022121304646134377,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 3.6606979370117188,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.02168990671634674,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 3.6801178455352783,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.020228713750839233,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 3.6707205772399902,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.018500352278351784,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 3.647545576095581,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.020223816856741905,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 3.644932985305786,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.019502121955156326,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 3.62888240814209,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.01926770992577076,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 3.675457715988159,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.01888471283018589,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 3.636434316635132,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.0201494712382555,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 3.679147720336914,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.02178201824426651,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 3.671905040740967,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.024073559790849686,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 3.6701040267944336,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.02135559171438217,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 3.648519992828369,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.019264448434114456,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 3.6985023021698,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.01922295056283474,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 3.651756763458252,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.01978331431746483,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 3.679884910583496,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.01968599483370781,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 3.6571812629699707,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.020741639658808708,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 3.6505298614501953,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.01842017099261284,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 3.6904940605163574,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.018004486337304115,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 3.7282047271728516,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.018821008503437042,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 3.668164014816284,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.018929854035377502,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 3.6672980785369873,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.01979168877005577,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 3.6684999465942383,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.01774384267628193,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 3.6577749252319336,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01876000687479973,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 3.6629252433776855,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.018197249621152878,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 3.688983678817749,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.019496265798807144,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 3.6602022647857666,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01825535111129284,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 3.6889238357543945,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.017113137990236282,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 3.684070587158203,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.017993126064538956,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 3.675426721572876,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.018354708328843117,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 3.675588607788086,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.016362832859158516,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 3.6709303855895996,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01772182621061802,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 3.6884102821350098,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.019271764904260635,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 3.6468968391418457,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.019040528684854507,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 3.690387010574341,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.019022610038518906,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 3.689873218536377,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.018533391878008842,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 3.6877284049987793,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.020825615152716637,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 3.662257671356201,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.021880341693758965,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 3.650700092315674,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.019974704831838608,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 3.684177875518799,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.01903127320110798,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 3.658148765563965,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.020999105647206306,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 3.68747615814209,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.02138398587703705,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 3.656355381011963,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.02216442860662937,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 3.6860342025756836,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.019359121099114418,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 3.654944658279419,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.020984770730137825,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 3.6885194778442383,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.02423972077667713,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 3.672712802886963,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.024544453248381615,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 3.6484768390655518,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.060352325439453,
+ "eval_runtime": 57.9556,
+ "eval_samples_per_second": 42.136,
+ "eval_steps_per_second": 1.329,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.024072550237178802,
+ "learning_rate": 0.000444677103403854,
+ "loss": 3.5796468257904053,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.028701433911919594,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 3.604318618774414,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.028119578957557678,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 3.608759641647339,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.025335192680358887,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 3.5760319232940674,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.0273929201066494,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 3.559441089630127,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.026567714288830757,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 3.626889228820801,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.027587885037064552,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 3.6226272583007812,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.02935059368610382,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 3.601921558380127,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.02982175536453724,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 3.59737491607666,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.028570258989930153,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 3.56337833404541,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.029890350997447968,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 3.6294307708740234,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.03206814453005791,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 3.590505599975586,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.03169345110654831,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 3.581584930419922,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.03313758969306946,
+ "learning_rate": 0.000441621628895411,
+ "loss": 3.6088438034057617,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.03317603841423988,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 3.6340456008911133,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.034738484770059586,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 3.5814096927642822,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.036428093910217285,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 3.5939526557922363,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.03293406963348389,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 3.577897787094116,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.031030720099806786,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 3.6077990531921387,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.03441236540675163,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 3.591923713684082,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.03500965237617493,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 3.6171927452087402,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.034443050622940063,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 3.619724750518799,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.03435051813721657,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 3.6443169116973877,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.03230702877044678,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 3.5912671089172363,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.03232988342642784,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 3.618985414505005,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.029298555105924606,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 3.6166677474975586,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.026723450049757957,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 3.61362886428833,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.028280874714255333,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 3.6410951614379883,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.026700228452682495,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 3.5804049968719482,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.023092713207006454,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 3.595339298248291,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.024123897776007652,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 3.647569417953491,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.02505572699010372,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 3.6276652812957764,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.024523509666323662,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 3.59525728225708,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.024351581931114197,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 3.6081674098968506,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.02530614472925663,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 3.6632542610168457,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.023438062518835068,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 3.6345431804656982,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.021853657439351082,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 3.598644256591797,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.022487422451376915,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 3.5966577529907227,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.021353935822844505,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 3.6328375339508057,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.021180452778935432,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 3.6285812854766846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.020491832867264748,
+ "learning_rate": 0.000435215814386134,
+ "loss": 3.592944622039795,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.021603770554065704,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 3.613507032394409,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.02146158367395401,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 3.6409506797790527,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.02102433145046234,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 3.6179330348968506,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.018684551119804382,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 3.6385130882263184,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.018416227772831917,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 3.642151355743408,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.01915144734084606,
+ "learning_rate": 0.000433781621332752,
+ "loss": 3.6089656352996826,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.018299665302038193,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 3.6063547134399414,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.018962757661938667,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 3.625917434692383,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.017437821254134178,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 3.600015640258789,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.017913147807121277,
+ "learning_rate": 0.000432823382695327,
+ "loss": 3.6428744792938232,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.018343618139624596,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 3.635901927947998,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.020372850820422173,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 3.620181083679199,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.02037600241601467,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 3.61129093170166,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.021827151998877525,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 3.5999555587768555,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.021806657314300537,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 3.642509937286377,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.02050699107348919,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 3.597583293914795,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.020436758175492287,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 3.6079912185668945,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.019341666251420975,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 3.6196017265319824,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.022348817437887192,
+ "learning_rate": 0.000430661245733797,
+ "loss": 3.619025468826294,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.019361121580004692,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 3.657182455062866,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.019258935004472733,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 3.6408963203430176,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.020622272044420242,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 3.6247220039367676,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.019510755315423012,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 3.6264584064483643,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.02078665979206562,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 3.6364388465881348,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01791800744831562,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 3.6029374599456787,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.01922667771577835,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 3.582937717437744,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.018473893404006958,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 3.6141819953918457,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.01846298575401306,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 3.638786554336548,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.01876973733305931,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 3.617130756378174,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.017995499074459076,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 3.645305633544922,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.02105429396033287,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 3.6444876194000244,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.049062252044678,
+ "eval_runtime": 58.0113,
+ "eval_samples_per_second": 42.095,
+ "eval_steps_per_second": 1.327,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.02128119394183159,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 3.524954080581665,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.028249500319361687,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 3.5311479568481445,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.03145391866564751,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 3.559330940246582,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.02994118444621563,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 3.5257620811462402,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.027448246255517006,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 3.5242764949798584,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.0296134352684021,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 3.5317091941833496,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.029576804488897324,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 3.5340428352355957,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.02941269613802433,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 3.5372023582458496,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.03126693144440651,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 3.5437135696411133,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.03335367515683174,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 3.5598273277282715,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.03028946928679943,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 3.5190839767456055,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.028629133477807045,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 3.5138821601867676,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.027070026844739914,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 3.530958414077759,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.028837760910391808,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 3.5233962535858154,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.029884297400712967,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 3.5639758110046387,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.024223247542977333,
+ "learning_rate": 0.000423881957237287,
+ "loss": 3.5555152893066406,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.026992836967110634,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 3.540196418762207,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.02818330004811287,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 3.5546071529388428,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.026401042938232422,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 3.5556247234344482,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.026020489633083344,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 3.568004608154297,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.02803782932460308,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 3.5828235149383545,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.02737978659570217,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 3.5642499923706055,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.02406371757388115,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 3.5530993938446045,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.02230692468583584,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 3.5573501586914062,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.021352406591176987,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 3.5526974201202393,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.02227999083697796,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 3.5866928100585938,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.024190247058868408,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 3.517117977142334,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.023196619004011154,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 3.4826502799987793,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.02388714998960495,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 3.5588626861572266,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.024692339822649956,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 3.54828143119812,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.023078206926584244,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 3.5679030418395996,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.020665735006332397,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 3.5500283241271973,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.02022322081029415,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 3.592104911804199,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.021137656643986702,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 3.583833694458008,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.021026931703090668,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 3.556403875350952,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.020423648878932,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 3.5839173793792725,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.020935002714395523,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 3.5794217586517334,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.020799759775400162,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 3.6033859252929688,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.019480884075164795,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 3.580097198486328,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.019300777465105057,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 3.550474166870117,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.020265772938728333,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 3.5711655616760254,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.019972780719399452,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 3.5636885166168213,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.019635623320937157,
+ "learning_rate": 0.000417272427439646,
+ "loss": 3.5851311683654785,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.021451901644468307,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 3.5956759452819824,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.019456414505839348,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 3.595731258392334,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.019757689908146858,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 3.569652557373047,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.020571604371070862,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 3.6007747650146484,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.02153848111629486,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 3.5915260314941406,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.020875850692391396,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 3.5582690238952637,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.02089419960975647,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 3.5756618976593018,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.01961098238825798,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 3.5918776988983154,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.021075716242194176,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 3.5342202186584473,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.020745476707816124,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 3.5976028442382812,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.02001352794468403,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 3.5686182975769043,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.021684160456061363,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 3.553445339202881,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.022551920264959335,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 3.5712709426879883,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.022758224979043007,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 3.6101667881011963,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.020714636892080307,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 3.5740506649017334,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.022694187238812447,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 3.5663979053497314,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.021664340049028397,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 3.5830888748168945,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.019072270020842552,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 3.5861973762512207,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.020153764635324478,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 3.566444158554077,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.019666558131575584,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 3.5789592266082764,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.02101624570786953,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 3.551722526550293,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.01951020397245884,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 3.6069014072418213,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.01985476352274418,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 3.5944924354553223,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.02195335365831852,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 3.5828046798706055,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.02105172537267208,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 3.5830953121185303,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.01888827420771122,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 3.5656847953796387,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.017867228016257286,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 3.5792243480682373,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.018577905371785164,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 3.56992506980896,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.020045632496476173,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 3.590876579284668,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.044339656829834,
+ "eval_runtime": 57.1435,
+ "eval_samples_per_second": 42.735,
+ "eval_steps_per_second": 1.347,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.024135645478963852,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 3.4856467247009277,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.031053155660629272,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 3.455773115158081,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.035190410912036896,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 3.506619930267334,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.03611094877123833,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 3.5227696895599365,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.034121073782444,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 3.4970145225524902,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.03629382699728012,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 3.506425619125366,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.03709045797586441,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 3.499274969100952,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.034826792776584625,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 3.5029163360595703,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.03344690054655075,
+ "learning_rate": 0.000407857329622967,
+ "loss": 3.4816911220550537,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.03136949986219406,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 3.5154659748077393,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.030533097684383392,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 3.509666919708252,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.03298318013548851,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 3.5096616744995117,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.031186232343316078,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 3.4752590656280518,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.028008371591567993,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 3.498826503753662,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.028206519782543182,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 3.5281882286071777,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.024417119100689888,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 3.506791830062866,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.025321070104837418,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 3.494256019592285,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.023430505767464638,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 3.533010482788086,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.023371374234557152,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 3.5197086334228516,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.02414027415215969,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 3.4707889556884766,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.026652060449123383,
+ "learning_rate": 0.000404858275823277,
+ "loss": 3.498159885406494,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.024486590176820755,
+ "learning_rate": 0.000404607816444578,
+ "loss": 3.50034236907959,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.023153992369771004,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 3.499504804611206,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.02222427912056446,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 3.4953579902648926,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.023033641278743744,
+ "learning_rate": 0.000403855947934478,
+ "loss": 3.5181496143341064,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.02141181193292141,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 3.552860736846924,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.023448696359992027,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 3.5153324604034424,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.022410666570067406,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 3.4963512420654297,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.020326802507042885,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 3.528319835662842,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.021117394790053368,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 3.5192837715148926,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.021317072212696075,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 3.5081021785736084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.021116144955158234,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 3.5167174339294434,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.019834019243717194,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 3.511960983276367,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.02089378982782364,
+ "learning_rate": 0.000401595976318565,
+ "loss": 3.493330955505371,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.021129710599780083,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 3.544156551361084,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.021163299679756165,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 3.5261003971099854,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.02105306088924408,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 3.543973445892334,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.02150859497487545,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 3.487217426300049,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.021666452288627625,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 3.4933671951293945,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.02180723287165165,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 3.55198073387146,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.02041625790297985,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 3.529571533203125,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.022977063432335854,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 3.526432752609253,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.022725161164999008,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 3.5258829593658447,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.021337240934371948,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 3.4897351264953613,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.020783131942152977,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 3.5497589111328125,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.022259317338466644,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 3.5225064754486084,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.020283272489905357,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 3.5453078746795654,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.020656265318393707,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 3.533703327178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.02104976586997509,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 3.4885146617889404,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.021411150693893433,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 3.533268690109253,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.021906673908233643,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 3.5483224391937256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.02415728010237217,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 3.5084290504455566,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.023435061797499657,
+ "learning_rate": 0.000396803974909638,
+ "loss": 3.5498428344726562,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.02183385379612446,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 3.484086751937866,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.022856134921312332,
+ "learning_rate": 0.000396297942393265,
+ "loss": 3.5339231491088867,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.021679896861314774,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 3.547191619873047,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.021262124180793762,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 3.5263822078704834,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.02265402302145958,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 3.5361135005950928,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.020080721005797386,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 3.5247459411621094,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.02187187783420086,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 3.54129695892334,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.02417500875890255,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 3.598796844482422,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.02416270412504673,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 3.5207936763763428,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.021958643570542336,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 3.5395452976226807,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.021813420578837395,
+ "learning_rate": 0.000394017098798633,
+ "loss": 3.5469255447387695,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.020607495680451393,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 3.554851531982422,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.02091527171432972,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 3.492576837539673,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.021089257672429085,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 3.5154709815979004,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.019659532234072685,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 3.5444283485412598,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.01896163448691368,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 3.589625358581543,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.020288823172450066,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 3.522411823272705,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.020222362130880356,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 3.5244359970092773,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.022349944338202477,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 3.5362396240234375,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.047028064727783,
+ "eval_runtime": 62.8948,
+ "eval_samples_per_second": 38.827,
+ "eval_steps_per_second": 1.224,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.02202913723886013,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 3.4249138832092285,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.03153969347476959,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 3.4436073303222656,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.03747870773077011,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 3.4879579544067383,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.036144088953733444,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 3.419151782989502,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.03983336314558983,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 3.4331436157226562,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.045912645757198334,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 3.4304769039154053,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.04005753993988037,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 3.478600025177002,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.03488151356577873,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 3.4650232791900635,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.03603382781147957,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 3.449807643890381,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.03817906603217125,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 3.4556589126586914,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.03523671627044678,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 3.477426767349243,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.03284068778157234,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 3.478672981262207,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.02992352657020092,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 3.4810047149658203,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.02858242578804493,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 3.467811107635498,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.027912171557545662,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 3.4326171875,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.028351375833153725,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 3.4533729553222656,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.027947887778282166,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 3.420295476913452,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.028510035946965218,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 3.438023328781128,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.027707934379577637,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 3.4844181537628174,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.0262764822691679,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 3.4363749027252197,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.02713681012392044,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 3.4501900672912598,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.0299485195428133,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 3.461942672729492,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.027047986164689064,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 3.454293727874756,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.02436736971139908,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 3.4663405418395996,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.02601189725100994,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 3.502030849456787,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.02463468722999096,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 3.462491035461426,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.021793190389871597,
+ "learning_rate": 0.00038509205478744,
+ "loss": 3.4476377964019775,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.023809107020497322,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 3.4701178073883057,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.02557182125747204,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 3.4932539463043213,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.023424552753567696,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 3.4779105186462402,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.02195148915052414,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 3.470506191253662,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.022897250950336456,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 3.48982572555542,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.023033805191516876,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 3.4845101833343506,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.022774485871195793,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 3.5060324668884277,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.02193611115217209,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 3.511518955230713,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.02130570448935032,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 3.468502998352051,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.020402755588293076,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 3.4786908626556396,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.022664198651909828,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 3.479928970336914,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.02454056777060032,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 3.479947566986084,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.022096732631325722,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 3.49478816986084,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.021499406546354294,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 3.494933605194092,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.022325703874230385,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 3.5041463375091553,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.020674455910921097,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 3.472404718399048,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.02131054364144802,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 3.497795343399048,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.02126218006014824,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 3.483281373977661,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.0217830128967762,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 3.4704298973083496,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.02156917378306389,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 3.4994399547576904,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.02047055773437023,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 3.4316084384918213,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.019766805693507195,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 3.452223777770996,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.020028885453939438,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 3.485278844833374,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.020312055945396423,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 3.4789156913757324,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.020725280046463013,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 3.495528221130371,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.02088077738881111,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 3.4972569942474365,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.021010037511587143,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 3.473562240600586,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.019641095772385597,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 3.4782915115356445,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.02063329517841339,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 3.4681782722473145,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.0213928259909153,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 3.472263813018799,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.02135022170841694,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 3.502300262451172,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.02053583413362503,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 3.489243745803833,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.021575238555669785,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 3.491074562072754,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.022335771471261978,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 3.5298843383789062,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.021198464557528496,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 3.5156898498535156,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.020986037328839302,
+ "learning_rate": 0.000375827577564042,
+ "loss": 3.4574475288391113,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.022639811038970947,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 3.4952611923217773,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.019711066037416458,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 3.531296730041504,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.02058061584830284,
+ "learning_rate": 0.000375051971546236,
+ "loss": 3.5340657234191895,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.019849255681037903,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 3.5028295516967773,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01954430714249611,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 3.51088809967041,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.019234873354434967,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 3.472560167312622,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.021085912361741066,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 3.468876838684082,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.020300406962633133,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 3.4903886318206787,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.02380351349711418,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 3.5069704055786133,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.048111915588379,
+ "eval_runtime": 58.312,
+ "eval_samples_per_second": 41.878,
+ "eval_steps_per_second": 1.32,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.02379327267408371,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 3.396376132965088,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.03204222396016121,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 3.4102582931518555,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.034993767738342285,
+ "learning_rate": 0.000372722041257003,
+ "loss": 3.385697364807129,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.03387003391981125,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 3.4120078086853027,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.03282478079199791,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 3.404122829437256,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.03053346462547779,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 3.368051767349243,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.029722390696406364,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 3.4066247940063477,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.0287226140499115,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 3.4212646484375,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.028156550601124763,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 3.4192593097686768,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.03021939843893051,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 3.3973188400268555,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.03314025700092316,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 3.422978401184082,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.03156686946749687,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 3.4052343368530273,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.03357478231191635,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 3.383730411529541,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.03261081501841545,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 3.421250820159912,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.03252755105495453,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 3.39884352684021,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.03177274018526077,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 3.4175422191619873,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.028845487162470818,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 3.3908143043518066,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.026045162230730057,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 3.4417710304260254,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.027937524020671844,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 3.410548210144043,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.025780271738767624,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 3.4069018363952637,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.026093080639839172,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 3.4199600219726562,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.025765180587768555,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 3.442903995513916,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.022773101925849915,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 3.394232749938965,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.025135623291134834,
+ "learning_rate": 0.000367268105959126,
+ "loss": 3.4466171264648438,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.02491890639066696,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 3.418689727783203,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.02521841786801815,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 3.4213061332702637,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.025186199694871902,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 3.437631130218506,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.022097431123256683,
+ "learning_rate": 0.000366226612810024,
+ "loss": 3.4014902114868164,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.024009618908166885,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 3.4292187690734863,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.022758696228265762,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 3.4105381965637207,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.02447321265935898,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 3.4196958541870117,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.025843804702162743,
+ "learning_rate": 0.000365184304613104,
+ "loss": 3.4343783855438232,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.022433951497077942,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 3.440580368041992,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.02250702679157257,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 3.4255261421203613,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.022953296080231667,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 3.442258358001709,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.021977661177515984,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 3.4058687686920166,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.021751178428530693,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 3.4211649894714355,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.022764205932617188,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 3.424694061279297,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.024827787652611732,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 3.458076238632202,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.021888602524995804,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 3.4509153366088867,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.021961702033877373,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 3.4147253036499023,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.02235778421163559,
+ "learning_rate": 0.000362575056487632,
+ "loss": 3.4392640590667725,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.021714260801672935,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 3.444026470184326,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.020996905863285065,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 3.4168434143066406,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.020528919994831085,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 3.4831130504608154,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.021056141704320908,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 3.430905342102051,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.021655110642313957,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 3.4588241577148438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.02344612590968609,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 3.4202680587768555,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.02264723740518093,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 3.4669015407562256,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.02377854660153389,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 3.5059216022491455,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.02299923449754715,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 3.4796524047851562,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.021110547706484795,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 3.4342408180236816,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.022060122340917587,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 3.426923990249634,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.021077830344438553,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 3.4636270999908447,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.02209082990884781,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 3.441979169845581,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.022408071905374527,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 3.490835666656494,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.022462666034698486,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 3.452770233154297,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.021295800805091858,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 3.4377408027648926,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.020457757636904716,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 3.4191250801086426,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.021163633093237877,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 3.4797091484069824,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.021495461463928223,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 3.4763145446777344,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.021050071343779564,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 3.4687209129333496,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.02150929905474186,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 3.4819254875183105,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.0212967898696661,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 3.490407705307007,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.02123691327869892,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 3.4418535232543945,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.02193271741271019,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 3.446892738342285,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.02065582573413849,
+ "learning_rate": 0.000356031397755195,
+ "loss": 3.4308691024780273,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.019487040117383003,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 3.487239122390747,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.021190891042351723,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 3.4635558128356934,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.0209872554987669,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 3.443793773651123,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.02035645768046379,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 3.473198652267456,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.024002201855182648,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 3.4244751930236816,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.057085990905762,
+ "eval_runtime": 62.4735,
+ "eval_samples_per_second": 39.089,
+ "eval_steps_per_second": 1.233,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.02450242079794407,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 3.361237049102783,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.030813416466116905,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 3.365708351135254,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.03279578685760498,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 3.325385570526123,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.03293970227241516,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 3.356811285018921,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.033190157264471054,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 3.406217098236084,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.033370476216077805,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 3.3579213619232178,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.03351330757141113,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 3.3755106925964355,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.034596871584653854,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 3.3576536178588867,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.03177076578140259,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 3.3123257160186768,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.0289607010781765,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 3.370959997177124,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.03265267238020897,
+ "learning_rate": 0.000351829234402338,
+ "loss": 3.332608938217163,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.03397805988788605,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 3.364213228225708,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.03100395016372204,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 3.388005256652832,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.03224753588438034,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 3.3267595767974854,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.03034576214849949,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 3.3954129219055176,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.026316817849874496,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 3.3543832302093506,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.027714375406503677,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 3.4190099239349365,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.028731998056173325,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 3.342031955718994,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.02845810353755951,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 3.356130599975586,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.02842753194272518,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 3.3709869384765625,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.02942183054983616,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 3.373588800430298,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.02958574891090393,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 3.377514600753784,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.029524607583880424,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 3.398663282394409,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.028092598542571068,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 3.3742518424987793,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.026652829721570015,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 3.387667179107666,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.024675756692886353,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 3.3737070560455322,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.02491692826151848,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 3.359217643737793,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.024601681157946587,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 3.3902688026428223,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.02313913404941559,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 3.3700339794158936,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.02418898604810238,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 3.4265692234039307,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.023382030427455902,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 3.378389835357666,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.02340470254421234,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 3.392937421798706,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.022166339680552483,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 3.39414119720459,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.022175129503011703,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 3.407447099685669,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.02368207648396492,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 3.415076971054077,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.02263982594013214,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 3.383986473083496,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.02147712931036949,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 3.403963565826416,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.022987166419625282,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 3.3883113861083984,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.02240179106593132,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 3.4349911212921143,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.02206592820584774,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 3.426473617553711,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.022650036960840225,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 3.4251773357391357,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.023399339988827705,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 3.371610641479492,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.02192695252597332,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 3.387253761291504,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.02241935208439827,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 3.434523105621338,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.023202672600746155,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 3.369771957397461,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.021333014592528343,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 3.439004421234131,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.023988714441657066,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 3.396838665008545,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.023885777220129967,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 3.4114322662353516,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.02263713628053665,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 3.4139351844787598,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.022404303774237633,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 3.4200823307037354,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.021705076098442078,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 3.4092609882354736,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.02273249812424183,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 3.3931846618652344,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.022552402690052986,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 3.396566867828369,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.023219924420118332,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 3.369025707244873,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.022560900077223778,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 3.4248225688934326,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.023370858281850815,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 3.431628942489624,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.02469109185039997,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 3.406843662261963,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.02417670376598835,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 3.445046901702881,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.0237257182598114,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 3.4490597248077393,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.02295520156621933,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 3.4009876251220703,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.023158475756645203,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 3.4383902549743652,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.023618478327989578,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 3.421168804168701,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.022055329754948616,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 3.382946729660034,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.021441346034407616,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 3.4464712142944336,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.023031961172819138,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 3.42844295501709,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.022832952439785004,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 3.398231029510498,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.022094765678048134,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 3.4270074367523193,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.0218376275151968,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 3.4681742191314697,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.02364739589393139,
+ "learning_rate": 0.000336518346307424,
+ "loss": 3.431920051574707,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.023402225226163864,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 3.4189023971557617,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.022775836288928986,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 3.409396171569824,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.02445383556187153,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 3.41017746925354,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.066352367401123,
+ "eval_runtime": 58.1658,
+ "eval_samples_per_second": 41.983,
+ "eval_steps_per_second": 1.324,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.02710847556591034,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 3.3155131340026855,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.03428799286484718,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 3.327408790588379,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.03434379771351814,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 3.3282251358032227,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.03405233472585678,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 3.2932448387145996,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.031194346025586128,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 3.336615800857544,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.031816478818655014,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 3.3375275135040283,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.03413299471139908,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 3.3710694313049316,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.034200165420770645,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 3.3248214721679688,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.030694160610437393,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 3.3297901153564453,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.031172804534435272,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 3.287217140197754,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.031914614140987396,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 3.3218531608581543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.0332055389881134,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 3.3454785346984863,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.034956980496644974,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 3.3193702697753906,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.03618055582046509,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 3.3661017417907715,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.030322827398777008,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 3.353989601135254,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.029768960550427437,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 3.3254497051239014,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.028036782518029213,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 3.3276920318603516,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.028774483129382133,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 3.3458826541900635,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.030403515323996544,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 3.3332793712615967,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.028468478471040726,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 3.3167755603790283,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.029126476496458054,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 3.299426794052124,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.030746638774871826,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 3.3581809997558594,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.030241351574659348,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 3.3415942192077637,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.029754415154457092,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 3.3525662422180176,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.026896746829152107,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 3.321808099746704,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.026361066848039627,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 3.329258441925049,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.027747981250286102,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 3.3774523735046387,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.026270100846886635,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 3.3419976234436035,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.02731695957481861,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 3.3068814277648926,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.026583299040794373,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 3.303804874420166,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.026615949347615242,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 3.3812599182128906,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.027221621945500374,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 3.327211618423462,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.025025667622685432,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 3.339163303375244,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.02593836560845375,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 3.3517372608184814,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.023697076365351677,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 3.379819869995117,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.02425902709364891,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 3.390653610229492,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.025133250281214714,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 3.364495277404785,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.025542905554175377,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 3.3584814071655273,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.026041125878691673,
+ "learning_rate": 0.000325372061241796,
+ "loss": 3.36777400970459,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.02391037344932556,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 3.3809385299682617,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.02379394695162773,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 3.354257583618164,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.0234722550958395,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 3.3736205101013184,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.024123650044202805,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 3.3589959144592285,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.024013696238398552,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 3.353874444961548,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.022617843002080917,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 3.364056348800659,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.022730547934770584,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 3.378067970275879,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.022257911041378975,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 3.379342555999756,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.023247094824910164,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 3.3537116050720215,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.022764021530747414,
+ "learning_rate": 0.000322712903929477,
+ "loss": 3.3969063758850098,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.02260403335094452,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 3.3727574348449707,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.021772351115942,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 3.335434675216675,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.02238459885120392,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 3.3440277576446533,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.0230540968477726,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 3.3997583389282227,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.0220821350812912,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 3.377815008163452,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.021497417241334915,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 3.3696370124816895,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.02163885347545147,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 3.3489794731140137,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.022394269704818726,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 3.3792812824249268,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.022390708327293396,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 3.3733272552490234,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.023712158203125,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 3.3716251850128174,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.02398860827088356,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 3.381779670715332,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.02309161052107811,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 3.3923277854919434,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.021400388330221176,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 3.407397747039795,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.022534485906362534,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 3.3740391731262207,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.023533202707767487,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 3.3877944946289062,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.023398835211992264,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 3.3936471939086914,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.020927201956510544,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 3.339545249938965,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.021453995257616043,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 3.3647522926330566,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.022847114130854607,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 3.405144691467285,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.022766996175050735,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 3.392810344696045,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.022103948518633842,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 3.3973004817962646,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.022282181307673454,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 3.409998893737793,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.025435522198677063,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 3.423318386077881,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.078423500061035,
+ "eval_runtime": 57.3189,
+ "eval_samples_per_second": 42.604,
+ "eval_steps_per_second": 1.343,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.02624760940670967,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 3.261288642883301,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.03552987053990364,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 3.2682106494903564,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.040380869060754776,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 3.2813782691955566,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.03861293941736221,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 3.258193016052246,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.04091731086373329,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 3.3052220344543457,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.043269723653793335,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 3.2820217609405518,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.04580320790410042,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 3.312005043029785,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.047336459159851074,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 3.3217713832855225,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.04587360844016075,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 3.2869949340820312,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.04002532362937927,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 3.2879223823547363,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.0368693508207798,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 3.2983577251434326,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.037317775189876556,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 3.3062524795532227,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.03655700013041496,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 3.31072735786438,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.033466216176748276,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 3.2954514026641846,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.030785970389842987,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 3.27952241897583,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.029669342562556267,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 3.2613229751586914,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.033505845814943314,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 3.2987258434295654,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.032395437359809875,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 3.296708583831787,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.031322430819272995,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 3.3160252571105957,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.029345186427235603,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 3.307143211364746,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.026715675368905067,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 3.286012649536133,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.028592750430107117,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 3.2528645992279053,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.028861764818429947,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 3.2795748710632324,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.02609528973698616,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 3.326317071914673,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.027718454599380493,
+ "learning_rate": 0.000309930621834911,
+ "loss": 3.268110513687134,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.026382867246866226,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 3.294499397277832,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.027083583176136017,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 3.3202130794525146,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.02516467310488224,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 3.3485021591186523,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.024148069322109222,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 3.3099541664123535,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.024552857503294945,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 3.2885942459106445,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.02551114559173584,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 3.3232040405273438,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.025423647835850716,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 3.3097832202911377,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.023211417719721794,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 3.3075966835021973,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.024623898789286613,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 3.2917377948760986,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.02677663415670395,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 3.356265068054199,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.02673642337322235,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 3.314903736114502,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.024466225877404213,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 3.30912446975708,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.025370748713612556,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 3.363640069961548,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.026536114513874054,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 3.304851531982422,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.026536893099546432,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 3.342902898788452,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.02645307406783104,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 3.3164432048797607,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.02523060515522957,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 3.34376859664917,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.02590431459248066,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 3.3479061126708984,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.0230655986815691,
+ "learning_rate": 0.000304866093757771,
+ "loss": 3.2890982627868652,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.02403893694281578,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 3.345456600189209,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.02238907851278782,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 3.3471767902374268,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.021377569064497948,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 3.319232225418091,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.02272353693842888,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 3.316892623901367,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.02413824200630188,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 3.3450968265533447,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.02334682270884514,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 3.3640642166137695,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.021966520696878433,
+ "learning_rate": 0.000303,
+ "loss": 3.358555316925049,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.023984894156455994,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 3.358398914337158,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.023520268499851227,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 3.3224587440490723,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.0231285709887743,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 3.3671562671661377,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.022818170487880707,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 3.361027717590332,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.02280694991350174,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 3.2923383712768555,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.021362345665693283,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 3.3346734046936035,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.02372218668460846,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 3.354748249053955,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.022949764505028725,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 3.3857789039611816,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.021954255178570747,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 3.343330144882202,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.022438595071434975,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 3.328618049621582,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.021969925612211227,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 3.355971574783325,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.022706981748342514,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 3.3529577255249023,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.024083172902464867,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 3.348170757293701,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.023259354755282402,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 3.3474369049072266,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.02218157798051834,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 3.369053840637207,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.02305331453680992,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 3.3622641563415527,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.022796178236603737,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 3.351510524749756,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.021299205720424652,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 3.339406728744507,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.022786924615502357,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 3.3182196617126465,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.021729128435254097,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 3.383904218673706,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.024604283273220062,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 3.2912864685058594,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.0894880294799805,
+ "eval_runtime": 58.3967,
+ "eval_samples_per_second": 41.817,
+ "eval_steps_per_second": 1.319,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.02673362009227276,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 3.270986795425415,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.04036599025130272,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 3.2358999252319336,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.047250088304281235,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 3.2706189155578613,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.04642819985747337,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 3.236340284347534,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.039405327290296555,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 3.203136682510376,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.038748081773519516,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 3.2636070251464844,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.039233580231666565,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 3.2751896381378174,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.03429320082068443,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 3.2765679359436035,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.03267038241028786,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 3.2387681007385254,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.03349406272172928,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 3.282162666320801,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.032002877444028854,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 3.246623992919922,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.031509071588516235,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 3.2374932765960693,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.02958357334136963,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 3.24362850189209,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.03248635679483414,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 3.237210273742676,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.03230655938386917,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 3.251129150390625,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.029688173905014992,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 3.2141575813293457,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.027773940935730934,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 3.264249086380005,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.02687031403183937,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 3.2513632774353027,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.02503623627126217,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 3.217961072921753,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.027071228250861168,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 3.2784645557403564,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.02600773423910141,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 3.2641382217407227,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.026998162269592285,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 3.2646782398223877,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.025579402223229408,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 3.2700581550598145,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.02645333670079708,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 3.2545266151428223,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.02717350795865059,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 3.2795705795288086,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.02675914391875267,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 3.2936904430389404,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.02751768007874489,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 3.2329440116882324,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.025446882471442223,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 3.274294376373291,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.02365768514573574,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 3.2585229873657227,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.0266884732991457,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 3.2872962951660156,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.02572469227015972,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 3.253828287124634,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.02540273219347,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 3.303926944732666,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.024694625288248062,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 3.264145851135254,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.027233170345425606,
+ "learning_rate": 0.000288343693342466,
+ "loss": 3.286924362182617,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.02635285072028637,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 3.2930331230163574,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.025157805532217026,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 3.2906906604766846,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.024729596450924873,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 3.277128219604492,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.023809000849723816,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 3.2957217693328857,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.02474421076476574,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 3.27791690826416,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.02439294196665287,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 3.2912981510162354,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.023601334542036057,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 3.317446231842041,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.0253264419734478,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 3.2742302417755127,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.02389134280383587,
+ "learning_rate": 0.000285947841605349,
+ "loss": 3.281032085418701,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.02451813407242298,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 3.3339295387268066,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.024324845522642136,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 3.290402889251709,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.024939008057117462,
+ "learning_rate": 0.000285149463934261,
+ "loss": 3.2933621406555176,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.024450505152344704,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 3.3055014610290527,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.02262270078063011,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 3.2964205741882324,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.02311946265399456,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 3.3102402687072754,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.024362705647945404,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 3.3020925521850586,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.024785032495856285,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 3.287290096282959,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.02358846180140972,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 3.334261894226074,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.024330584332346916,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 3.2799081802368164,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.023546984419226646,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 3.3228847980499268,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.023060832172632217,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 3.325793981552124,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.02388455718755722,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 3.2892062664031982,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.023740701377391815,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 3.2969369888305664,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.02458217367529869,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 3.315610408782959,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.023173579946160316,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 3.2561936378479004,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.024137936532497406,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 3.2839202880859375,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.022436272352933884,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 3.3057351112365723,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.023997878655791283,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 3.3583154678344727,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.024119896814227104,
+ "learning_rate": 0.000280627938758204,
+ "loss": 3.322892427444458,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.022283365949988365,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 3.3145735263824463,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.023433363065123558,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 3.3564529418945312,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.02224746160209179,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 3.2985916137695312,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.023786570876836777,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 3.302539587020874,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.025200961157679558,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 3.309976577758789,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.023849809542298317,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 3.343660354614258,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.023311031982302666,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 3.2978909015655518,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.023174891248345375,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 3.326188802719116,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.02681034244596958,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 3.3422017097473145,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.109009265899658,
+ "eval_runtime": 58.0506,
+ "eval_samples_per_second": 42.067,
+ "eval_steps_per_second": 1.326,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.028057295829057693,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 3.1825037002563477,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.034786805510520935,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 3.200389862060547,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.03751441463828087,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 3.2047512531280518,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.035590436309576035,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 3.1735124588012695,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.03426426649093628,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 3.199483871459961,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.037030890583992004,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 3.213322639465332,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.032468877732753754,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 3.179354190826416,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.027202758938074112,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 3.216036319732666,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.031184563413262367,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 3.208205223083496,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.029414812102913857,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 3.180370569229126,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.02679901383817196,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 3.210325002670288,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.02863631211221218,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 3.1601076126098633,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.029135366901755333,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 3.2262887954711914,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.030535712838172913,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 3.23160719871521,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.030563877895474434,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 3.255211353302002,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.02906505949795246,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 3.2006797790527344,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.030871668830513954,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 3.2442867755889893,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.02913421392440796,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 3.218442440032959,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.030087953433394432,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 3.2164535522460938,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.034147944301366806,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 3.218660831451416,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.03593966364860535,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 3.2240614891052246,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.032090652734041214,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 3.2309441566467285,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.02776424027979374,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 3.2702016830444336,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.027174271643161774,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 3.19810152053833,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.028479408472776413,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 3.2178421020507812,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.026820644736289978,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 3.2634482383728027,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.026427917182445526,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 3.2452914714813232,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.02679610252380371,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 3.243825912475586,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.026115626096725464,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 3.218787670135498,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.026686571538448334,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 3.2785303592681885,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.02685859426856041,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 3.2572226524353027,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.026968171820044518,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 3.2625553607940674,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.025397246703505516,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 3.262303352355957,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.025660471990704536,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 3.269401788711548,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.02713891491293907,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 3.2416529655456543,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.027591092512011528,
+ "learning_rate": 0.00026868712586269,
+ "loss": 3.2814316749572754,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.024462848901748657,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 3.249725103378296,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.024868478998541832,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 3.231574535369873,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.027015401050448418,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 3.282627582550049,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.02678140625357628,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 3.2208900451660156,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.026830527931451797,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 3.260575294494629,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.02430890128016472,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 3.2522997856140137,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.027047613635659218,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 3.2721078395843506,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.026861369609832764,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 3.2235946655273438,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.026650549843907356,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 3.302743434906006,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.02551819011569023,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 3.265763282775879,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.024843450635671616,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 3.272397518157959,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.024423206225037575,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 3.253495454788208,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.024180661886930466,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 3.260438919067383,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.023554982617497444,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 3.2342851161956787,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.023470086976885796,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 3.251622438430786,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.024405231699347496,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 3.2588815689086914,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.024147814139723778,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 3.2723565101623535,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.023217910900712013,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 3.2555792331695557,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.024921158328652382,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 3.26884126663208,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.02455919235944748,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 3.2674927711486816,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.024725932627916336,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 3.2511110305786133,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.025769414380192757,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 3.290651321411133,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.025928806513547897,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 3.302396297454834,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.02499072253704071,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 3.2813124656677246,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.025431277230381966,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 3.234684705734253,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.026871860027313232,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 3.2800331115722656,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.02578776702284813,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 3.3117146492004395,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.026318378746509552,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 3.257615089416504,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.025542527437210083,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 3.2969868183135986,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.025723429396748543,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 3.290259838104248,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.024588774889707565,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 3.284090280532837,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.024619122967123985,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 3.284879207611084,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.024441618472337723,
+ "learning_rate": 0.00025996500713765,
+ "loss": 3.283104658126831,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.02402760274708271,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 3.3013617992401123,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.025028521195054054,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 3.2892568111419678,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.029445001855492592,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 3.2474446296691895,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.128176689147949,
+ "eval_runtime": 58.8588,
+ "eval_samples_per_second": 41.489,
+ "eval_steps_per_second": 1.308,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.0298826452344656,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 3.138103485107422,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.0387548953294754,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 3.161492347717285,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.03816070780158043,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 3.1739673614501953,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.03409738838672638,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 3.182478189468384,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.033135563135147095,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 3.1720199584960938,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.03203624114394188,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 3.1699862480163574,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.031237754970788956,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 3.179363250732422,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.03342146426439285,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 3.1822280883789062,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.03561089560389519,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 3.1767094135284424,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.034710463136434555,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 3.218817949295044,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.030870331451296806,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 3.168602705001831,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.028004391118884087,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 3.224155902862549,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.030953701585531235,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 3.1789557933807373,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.030289432033896446,
+ "learning_rate": 0.000255486047794226,
+ "loss": 3.1890387535095215,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.030684132128953934,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 3.195427417755127,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.03022196516394615,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 3.1984965801239014,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.026935536414384842,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 3.2014272212982178,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.02702922560274601,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 3.205538034439087,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.026682600378990173,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 3.1954591274261475,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.0273384228348732,
+ "learning_rate": 0.000253907826333243,
+ "loss": 3.1782338619232178,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.028587577864527702,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 3.199565887451172,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.028848132118582726,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 3.181786298751831,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.027535341680049896,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 3.2013628482818604,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.02756025455892086,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 3.1749587059020996,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.027358995750546455,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 3.161881923675537,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.025918981060385704,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 3.225224256515503,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.026479976251721382,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 3.2129316329956055,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.02685079537332058,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 3.1651554107666016,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.027378706261515617,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 3.193108558654785,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.027240974828600883,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 3.190605640411377,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.028065402060747147,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 3.2068259716033936,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.028329014778137207,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 3.2316060066223145,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.028920065611600876,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 3.1933932304382324,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.028299354016780853,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 3.192711353302002,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.025996020063757896,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 3.243171453475952,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.0264595914632082,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 3.199758529663086,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.02704528719186783,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 3.2390713691711426,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.026571933180093765,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 3.219440460205078,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.025562983006238937,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 3.2538228034973145,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.024683184921741486,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 3.2372591495513916,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.025656528770923615,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 3.250558853149414,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.024807540699839592,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 3.1856563091278076,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.026149194687604904,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 3.172186851501465,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.02617705799639225,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 3.250445604324341,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.026542512699961662,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 3.220273494720459,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.024963509291410446,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 3.2330946922302246,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.027144264429807663,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 3.1860971450805664,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.02665683813393116,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 3.257209062576294,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.024872658774256706,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 3.2266182899475098,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.025235185399651527,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 3.2134504318237305,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.025029495358467102,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 3.2150087356567383,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.025630896911025047,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 3.282879114151001,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.02722341939806938,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 3.2366206645965576,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.02522372268140316,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 3.2840523719787598,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.02494790591299534,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 3.2405319213867188,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.026662737131118774,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 3.239692211151123,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.026082608848810196,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 3.2156195640563965,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.02423974685370922,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 3.1863760948181152,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.025226688012480736,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 3.20874285697937,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.02428225241601467,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 3.2209601402282715,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.023819707334041595,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 3.2564258575439453,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.025661710649728775,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 3.244133949279785,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.02423049882054329,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 3.2376630306243896,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.02382473088800907,
+ "learning_rate": 0.000242380656502223,
+ "loss": 3.2951290607452393,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.02488875202834606,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 3.2848310470581055,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.02575223706662655,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 3.2574503421783447,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.02462848648428917,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 3.2464442253112793,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.024912388995289803,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 3.2459192276000977,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.025020884349942207,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 3.213622808456421,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.026330886408686638,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 3.2305431365966797,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.02708602324128151,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 3.2487688064575195,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.030908891931176186,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 3.2110159397125244,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.146673202514648,
+ "eval_runtime": 58.7379,
+ "eval_samples_per_second": 41.574,
+ "eval_steps_per_second": 1.311,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.02919941022992134,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 3.1533267498016357,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.03379782289266586,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 3.1436538696289062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.033698830753564835,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 3.15297532081604,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.03077736124396324,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 3.128848075866699,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.031112324446439743,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 3.1461119651794434,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.03346852958202362,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 3.113943099975586,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.03307225927710533,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 3.147498607635498,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.030230754986405373,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 3.1162924766540527,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.02982981689274311,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 3.144481658935547,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.031735632568597794,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 3.133862018585205,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.0312766470015049,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 3.110287666320801,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.029902562499046326,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 3.1384904384613037,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.0306564774364233,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 3.1138200759887695,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.03084711916744709,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 3.155240535736084,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.03183220326900482,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 3.1593515872955322,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.031826410442590714,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 3.1517653465270996,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.030141707509756088,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 3.1291589736938477,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.03158571943640709,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 3.178101062774658,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.03510458394885063,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 3.138767719268799,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.03660859912633896,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 3.1708762645721436,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.034004807472229004,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 3.1404073238372803,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.03077496401965618,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 3.159081220626831,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.030384650453925133,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 3.1872847080230713,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.02957277186214924,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 3.165973424911499,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.029529055580496788,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 3.123840808868408,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.027776138857007027,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 3.1705269813537598,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.028943249955773354,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 3.1181654930114746,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.028178401291370392,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 3.201678991317749,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.02865665778517723,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 3.18479061126709,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.029173649847507477,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 3.1844916343688965,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.0280170775949955,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 3.187105417251587,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.029103878885507584,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 3.183281660079956,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.029045486822724342,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 3.1848196983337402,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.028364380821585655,
+ "learning_rate": 0.000231465389734324,
+ "loss": 3.197348117828369,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.029651399701833725,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 3.1429920196533203,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.0281438659876585,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 3.1916513442993164,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.029437677934765816,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 3.172391176223755,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.03433044254779816,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 3.2296864986419678,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.03259492665529251,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 3.1962363719940186,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.0299630519002676,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 3.169673442840576,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.028495769947767258,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 3.1855409145355225,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.0290420800447464,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 3.23667573928833,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.03340664133429527,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 3.1873788833618164,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.03253626823425293,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 3.204796552658081,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.028142794966697693,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 3.193922281265259,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.028805622830986977,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 3.1863901615142822,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.030033612623810768,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 3.198482036590576,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.028154220432043076,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 3.215069055557251,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.028098132461309433,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 3.2197628021240234,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.028679659590125084,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 3.2062196731567383,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.029104258865118027,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 3.2182908058166504,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.029629778116941452,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 3.2165639400482178,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.028304366394877434,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 3.1783785820007324,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.026900488883256912,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 3.1876602172851562,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.027047287672758102,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 3.1879465579986572,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.027424108237028122,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 3.2407851219177246,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.029319508001208305,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 3.1980130672454834,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.027114372700452805,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 3.211357593536377,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.02579840086400509,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 3.2067861557006836,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.02767331339418888,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 3.198587417602539,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.026645265519618988,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 3.177441120147705,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.026181906461715698,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 3.226815700531006,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.027218692004680634,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 3.1731510162353516,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.027551919221878052,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 3.2268621921539307,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.025782009586691856,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 3.2052621841430664,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.024231620132923126,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 3.201770544052124,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.02474607340991497,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 3.1935834884643555,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.026299327611923218,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 3.223275899887085,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.02439914271235466,
+ "learning_rate": 0.00022244633283095,
+ "loss": 3.2513954639434814,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.02650653012096882,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 3.1914851665496826,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.024847298860549927,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 3.2173104286193848,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.028475770726799965,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 3.2184863090515137,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.168707847595215,
+ "eval_runtime": 57.4035,
+ "eval_samples_per_second": 42.541,
+ "eval_steps_per_second": 1.341,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.02837727963924408,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 3.119239330291748,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.03562312573194504,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 3.092813014984131,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.0344194732606411,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 3.0802855491638184,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.03188958019018173,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 3.1177961826324463,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.032181013375520706,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 3.1343281269073486,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.031085725873708725,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 3.118828296661377,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.031995322555303574,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 3.110109329223633,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.02949182130396366,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 3.0972743034362793,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.029299668967723846,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 3.1257238388061523,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.030558617785573006,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 3.1216022968292236,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.03181575983762741,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 3.093721866607666,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.030934303998947144,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 3.0910396575927734,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.030235569924116135,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 3.133371114730835,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.03350163996219635,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 3.1178882122039795,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.03468719869852066,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 3.1222469806671143,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.034203238785266876,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 3.12453031539917,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.03321516141295433,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 3.115428924560547,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.03298434242606163,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 3.127288341522217,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.029987169429659843,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 3.1194515228271484,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.031271349638700485,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 3.1284165382385254,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.032201699912548065,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 3.1912736892700195,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.0299740731716156,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 3.1233785152435303,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.029613154008984566,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 3.1535067558288574,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.03303765505552292,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 3.1550910472869873,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.032644059509038925,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 3.1367573738098145,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.031695179641246796,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 3.1350858211517334,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.030685821548104286,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 3.121593952178955,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.03091946430504322,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 3.113060235977173,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.0282905176281929,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 3.184551239013672,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.030020160600543022,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 3.1673004627227783,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.031114062294363976,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 3.162588596343994,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.03007666766643524,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 3.132836103439331,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.029013793915510178,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 3.1552720069885254,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.02784561738371849,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 3.189272165298462,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.02759280428290367,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 3.141814708709717,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.02699868381023407,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 3.1127257347106934,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.02782190404832363,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 3.193972587585449,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.027298318222165108,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 3.1393446922302246,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.027321161702275276,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 3.0926666259765625,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.026399722322821617,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 3.1578259468078613,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.02770969271659851,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 3.1816914081573486,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.028344258666038513,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 3.1860158443450928,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.028091782703995705,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 3.1585092544555664,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.026564039289951324,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 3.131692409515381,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.026349127292633057,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 3.160431385040283,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.030081138014793396,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 3.1647543907165527,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.029250919818878174,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 3.1522226333618164,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.0282477755099535,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 3.139249324798584,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.0284534003585577,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 3.154621124267578,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.02677394263446331,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 3.1375081539154053,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.026511570438742638,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 3.1355676651000977,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.02766270563006401,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 3.143874168395996,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.027013977989554405,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 3.1602487564086914,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.026629449799656868,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 3.1549456119537354,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.0272684209048748,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 3.1739020347595215,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.027645915746688843,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 3.172959327697754,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.025413552299141884,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 3.1649041175842285,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.028406929224729538,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 3.1821887493133545,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.027753977105021477,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 3.163238048553467,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.02693209983408451,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 3.1921019554138184,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.02671878971159458,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 3.159438133239746,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.027099428698420525,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 3.1806793212890625,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.025098824873566628,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 3.158310651779175,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.02655479498207569,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 3.1927309036254883,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.025836164131760597,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 3.1578540802001953,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.02657160349190235,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 3.1658592224121094,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.027223270386457443,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 3.1760241985321045,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.025695348158478737,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 3.202389717102051,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.025819813832640648,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 3.2111103534698486,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.02621106244623661,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 3.1560299396514893,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.026585502550005913,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 3.176602840423584,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.03038652054965496,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 3.1799943447113037,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.193845748901367,
+ "eval_runtime": 57.9911,
+ "eval_samples_per_second": 42.11,
+ "eval_steps_per_second": 1.328,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.03135736659169197,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 3.070986270904541,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.038986921310424805,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 3.0708835124969482,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.035115309059619904,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 3.1070001125335693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.032868221402168274,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 3.082258701324463,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.03703184798359871,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 3.057744264602661,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.03831101953983307,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 3.0789475440979004,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.036301176995038986,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 3.0565457344055176,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.03166409209370613,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 3.0574910640716553,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.02984458953142166,
+ "learning_rate": 0.000201141724176723,
+ "loss": 3.080018997192383,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.034001126885414124,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 3.0813302993774414,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.03814122825860977,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 3.0579185485839844,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.03870106488466263,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 3.0857529640197754,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.04005642607808113,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 3.0880727767944336,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.03760755434632301,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 3.0795767307281494,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.03260907158255577,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 3.1006722450256348,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.029866183176636696,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 3.0850343704223633,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.030988436192274094,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 3.0743892192840576,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.03698134422302246,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 3.056002378463745,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.03696191683411598,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 3.069124698638916,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.03453022986650467,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 3.1017374992370605,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.032458193600177765,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 3.1166419982910156,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.02892368845641613,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 3.0863287448883057,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.030625877901911736,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 3.0938801765441895,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.030280394479632378,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 3.086432933807373,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.029181072488427162,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 3.0977859497070312,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.028039127588272095,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 3.112367868423462,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.02910272404551506,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 3.1253533363342285,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.03086872771382332,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 3.103689670562744,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.02883610874414444,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 3.111599922180176,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.028490116819739342,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 3.1410303115844727,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.028714407235383987,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 3.133023738861084,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.029553238302469254,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 3.107003688812256,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.02997349575161934,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 3.135680675506592,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.029087591916322708,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 3.1103882789611816,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.029947761446237564,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 3.0961623191833496,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.02950555644929409,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 3.1162495613098145,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.029569562524557114,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 3.096372127532959,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.031366728246212006,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 3.120089530944824,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.0313473716378212,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 3.100785732269287,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.028021851554512978,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 3.0940558910369873,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.027477117255330086,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 3.1210484504699707,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.029690749943256378,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 3.1350059509277344,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.029659733176231384,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 3.1220788955688477,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.028260033577680588,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 3.1000189781188965,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.027554461732506752,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 3.1044349670410156,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.027929967269301414,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 3.1353745460510254,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.029309319332242012,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 3.1578292846679688,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.029183287173509598,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 3.1217150688171387,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.02707512304186821,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 3.119617223739624,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.03006536327302456,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 3.1481480598449707,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.03026784397661686,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 3.142770767211914,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.02827967144548893,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 3.1582541465759277,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.028892774134874344,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 3.1530709266662598,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.03023783676326275,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 3.1102521419525146,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.027951840311288834,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 3.139113426208496,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.02807226963341236,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 3.149428367614746,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.029721451923251152,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 3.120518207550049,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.02867680788040161,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 3.1109931468963623,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.02663845382630825,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 3.116053581237793,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.028366081416606903,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 3.1444268226623535,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.028931576758623123,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 3.163870096206665,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.02664567157626152,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 3.1349146366119385,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.028031354770064354,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 3.1638219356536865,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.029421644285321236,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 3.1601340770721436,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.027983693405985832,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 3.1636788845062256,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.02719871513545513,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 3.1246869564056396,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.03030397929251194,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 3.1749579906463623,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.029340893030166626,
+ "learning_rate": 0.000186516746349841,
+ "loss": 3.1751790046691895,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.027624106034636497,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 3.178704023361206,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.02830546908080578,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 3.177117347717285,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.03168068826198578,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 3.125427722930908,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.033327918499708176,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 3.13071870803833,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.218388080596924,
+ "eval_runtime": 57.611,
+ "eval_samples_per_second": 42.388,
+ "eval_steps_per_second": 1.337,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.03336749225854874,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 3.064338445663452,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.038479577749967575,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 3.0478687286376953,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.032562077045440674,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 3.019803762435913,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.0331728458404541,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 3.0599141120910645,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.03361169621348381,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 3.0315823554992676,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.032973092049360275,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 3.025928258895874,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.030388766899704933,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 3.070671558380127,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.030899515375494957,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 3.012938976287842,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.0335412323474884,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 3.0156331062316895,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.035731345415115356,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 3.0578935146331787,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.03379053995013237,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 3.048701286315918,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.031242679804563522,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 3.0832977294921875,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.030997611582279205,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 3.0663583278656006,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.03397331386804581,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 3.0115303993225098,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.03349050134420395,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 3.02864933013916,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.031079065054655075,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 3.053401470184326,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.029346197843551636,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 3.0700926780700684,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.030784208327531815,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 3.066525936126709,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.03368140384554863,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 3.0956242084503174,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.03356792405247688,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 3.0394320487976074,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.029909325763583183,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 3.031525135040283,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.030711578205227852,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 3.075474977493286,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.03452072665095329,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 3.0228147506713867,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.03235075995326042,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 3.0664501190185547,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.029671598225831985,
+ "learning_rate": 0.000179445406945268,
+ "loss": 3.055302143096924,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.033868469297885895,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 3.053016185760498,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.0329158715903759,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 3.1017332077026367,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.031983695924282074,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 3.0706472396850586,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.02963697351515293,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 3.0711426734924316,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.03162391856312752,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 3.044867515563965,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.030773654580116272,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 3.0963196754455566,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.031220417469739914,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 3.0892245769500732,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.030928561463952065,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 3.1181323528289795,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.03192073106765747,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 3.066584348678589,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.03075495734810829,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 3.0436511039733887,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.02895325981080532,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 3.123373031616211,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.029738714918494225,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 3.1049060821533203,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.031937118619680405,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 3.099111318588257,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.029886579141020775,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 3.0970301628112793,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.029569782316684723,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 3.062499761581421,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.02924283780157566,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 3.0618538856506348,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.030955800786614418,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 3.0745606422424316,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.03145027533173561,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 3.122938632965088,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.028926467522978783,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 3.0808773040771484,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.029592076316475868,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 3.1167049407958984,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.028070243075489998,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 3.0781636238098145,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.028642063960433006,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 3.093937397003174,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.0290757454931736,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 3.1079559326171875,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.027867283672094345,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 3.0885753631591797,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.02896495908498764,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 3.0889246463775635,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.02777806483209133,
+ "learning_rate": 0.000173176617304673,
+ "loss": 3.084439277648926,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.02904277853667736,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 3.125246047973633,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.029010627418756485,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 3.111422061920166,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.028178894892334938,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 3.115067481994629,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.02763799950480461,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 3.1228506565093994,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.028052521869540215,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 3.1150546073913574,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.029438253492116928,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 3.0921788215637207,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.028195543214678764,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 3.1408450603485107,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.02937602624297142,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 3.111612319946289,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.028430236503481865,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 3.093441963195801,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.027624569833278656,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 3.1184372901916504,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.029797039926052094,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 3.1074559688568115,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.028728218749165535,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 3.161461114883423,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.02704412117600441,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 3.13325834274292,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.028067929670214653,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 3.093172550201416,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.028848597779870033,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 3.1461825370788574,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.02956785447895527,
+ "learning_rate": 0.00016935382741164,
+ "loss": 3.141045570373535,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.028214773163199425,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 3.136033773422241,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.029155246913433075,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 3.0986382961273193,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.02968468703329563,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 3.11630916595459,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.028356682509183884,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 3.1134822368621826,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.03301718086004257,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 3.118591785430908,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.246211528778076,
+ "eval_runtime": 57.5155,
+ "eval_samples_per_second": 42.458,
+ "eval_steps_per_second": 1.339,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.033637888729572296,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 3.0192506313323975,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.04057888314127922,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 2.977440595626831,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.03579854592680931,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 2.9902219772338867,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.03131850063800812,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 2.9988656044006348,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.03262636810541153,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 3.037747621536255,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.03260403499007225,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 2.987508773803711,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.033165134489536285,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 3.0511183738708496,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.03303489834070206,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 3.0083041191101074,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.03185006603598595,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 2.9966228008270264,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.03135651350021362,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 3.036759376525879,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.03216925635933876,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 3.0304150581359863,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.03231671452522278,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 3.0300474166870117,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.03263348713517189,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 3.0298163890838623,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.032594650983810425,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 2.98789119720459,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.03406650573015213,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 3.0102760791778564,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.03379780799150467,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 3.0564732551574707,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.03172587603330612,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 3.0410943031311035,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.02938298135995865,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 3.0517544746398926,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.03206252679228783,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 3.0352776050567627,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.033386752009391785,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 3.020535945892334,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.033984702080488205,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 3.0552730560302734,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0329931266605854,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 3.0075573921203613,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.032101184129714966,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 3.046388864517212,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.032873060554265976,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 3.0340287685394287,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.030672110617160797,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 3.0347838401794434,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.030014054849743843,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 3.0315332412719727,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.02932834066450596,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 3.084340810775757,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.029181934893131256,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 3.032886028289795,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.02966221794486046,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 3.032766819000244,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.029713116586208344,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 3.028778076171875,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.029243769124150276,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 3.0492146015167236,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.02997511625289917,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 3.0772786140441895,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.02930292673408985,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 3.055750608444214,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.03037521243095398,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 3.033710479736328,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.029957240447402,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 3.0394721031188965,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.02925589308142662,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 3.0878353118896484,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.03003518097102642,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 3.048628330230713,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.031545430421829224,
+ "learning_rate": 0.000159218843594243,
+ "loss": 3.0775623321533203,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.031310439109802246,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 3.01645565032959,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.03074057213962078,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 3.0651440620422363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.028704499825835228,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 3.0527148246765137,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.029061976820230484,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 3.0854501724243164,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.03034680150449276,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 3.056356906890869,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.028715113177895546,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 3.0874247550964355,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.030256496742367744,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 3.052424430847168,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.02949833683669567,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 3.0663650035858154,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.028980428352952003,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 3.0700113773345947,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.028526026755571365,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 3.0856711864471436,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.029572727158665657,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 3.0862035751342773,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.02955584041774273,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 3.030802011489868,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.02881692908704281,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 3.074392318725586,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.029564999043941498,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 3.057894229888916,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.030315004289150238,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 3.096233367919922,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.030666006729006767,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 3.040799617767334,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.028901133686304092,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 3.055051803588867,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.030359387397766113,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 3.085972309112549,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.03117106668651104,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 3.0786352157592773,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.029238386079669,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 3.0375771522521973,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.029324093833565712,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 3.1140198707580566,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.02875387854874134,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 3.090881109237671,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.029023943468928337,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 3.0755701065063477,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.0291463453322649,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 3.0723776817321777,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.02795177884399891,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 3.058993339538574,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.029499078169465065,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 3.0856404304504395,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.029889609664678574,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 3.0622942447662354,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.03107610158622265,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 3.114957332611084,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.02967257983982563,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 3.088871479034424,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.029068559408187866,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 3.1103081703186035,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.029334546998143196,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 3.0841622352600098,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.029259610921144485,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 3.0662848949432373,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.030072513967752457,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 3.0900328159332275,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.033321622759103775,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 3.0871129035949707,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.271463394165039,
+ "eval_runtime": 57.9048,
+ "eval_samples_per_second": 42.173,
+ "eval_steps_per_second": 1.33,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.03635408729314804,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 2.9576284885406494,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.03927835449576378,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 2.999624729156494,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.03371119126677513,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 2.971052646636963,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.033049341291189194,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 2.980811595916748,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.03830729424953461,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 2.9985947608947754,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.036573827266693115,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 2.9892568588256836,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.031196987256407738,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 2.9854254722595215,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.03376495838165283,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 2.9840264320373535,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.036272983998060226,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 3.0168871879577637,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.03328242897987366,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 2.9815053939819336,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.03158426657319069,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 2.994525909423828,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.03320910036563873,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 3.008638858795166,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.03160254284739494,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 2.979243040084839,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.03158991411328316,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 2.938634157180786,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.031116485595703125,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 2.9957656860351562,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.030757686123251915,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 3.0106635093688965,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.02992827817797661,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 2.996793746948242,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.0330309197306633,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 2.9859161376953125,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.029488038271665573,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 3.0081536769866943,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.03095114603638649,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 3.006385087966919,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.03194110840559006,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 3.0125250816345215,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.03306785598397255,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 2.9640283584594727,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.030883541330695152,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 3.0223653316497803,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.03224025294184685,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 3.0274910926818848,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.03182433918118477,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 3.047736406326294,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.03387221321463585,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 2.9821343421936035,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.03407924994826317,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 3.041207790374756,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.031546324491500854,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 3.0298609733581543,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.03354770317673683,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 3.0212411880493164,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.031692177057266235,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 3.007350444793701,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.03218817710876465,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 3.0019829273223877,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.03233771771192551,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 3.0400614738464355,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.028998777270317078,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 3.022221088409424,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.030422436073422432,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 3.0071797370910645,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.030207227915525436,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 2.9762964248657227,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.0305719543248415,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 3.075176239013672,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.03173545375466347,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 3.028982162475586,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.02799576334655285,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 3.0242385864257812,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.03140150383114815,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 3.012221336364746,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.029416466131806374,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 3.0037949085235596,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.028385788202285767,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 3.0350992679595947,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.031597770750522614,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 3.065139055252075,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.02926146425306797,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 3.014230728149414,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.02970231883227825,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 3.06636381149292,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.029684338718652725,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 3.024610996246338,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.02908361330628395,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 3.0093719959259033,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.029891187325119972,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 3.0735411643981934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.028607673943042755,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 3.044274091720581,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.029806701466441154,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 3.0449328422546387,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.03161340951919556,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 3.046813488006592,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.028860243037343025,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 3.011828660964966,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.0285879448056221,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 3.043640613555908,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.0296401996165514,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 3.071589469909668,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.02807026542723179,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 3.018320083618164,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.029037052765488625,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 3.0471763610839844,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.029226379469037056,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 3.037463426589966,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.029322804883122444,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 3.0472755432128906,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.028253905475139618,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 3.0265414714813232,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.028875533491373062,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 3.063941717147827,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.028460513800382614,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 3.0512683391571045,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.02951696142554283,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 3.0528573989868164,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.030593490228056908,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 3.057919979095459,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.028074532747268677,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 3.0263280868530273,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.02955678664147854,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 3.081322193145752,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.029724644497036934,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 3.059583902359009,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.029713226482272148,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 3.018678903579712,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.029758386313915253,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 3.0379555225372314,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.028971998021006584,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 3.041267156600952,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.02968725562095642,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 3.0675506591796875,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.029876509681344032,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 3.051586151123047,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.02853119559586048,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 3.0335168838500977,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.03374241292476654,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 3.027909278869629,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.030695006251335144,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 2.9685583114624023,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.0339728482067585,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 2.936551570892334,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.03372679278254509,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 2.9631872177124023,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.03172554448246956,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 2.969510078430176,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.030516354367136955,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 2.930342674255371,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.03261839970946312,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 2.9920477867126465,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.030618635937571526,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 2.9693596363067627,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.030576011165976524,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 2.929804801940918,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.03224785998463631,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 2.95280122756958,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.030605586245656013,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 2.931358575820923,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.031206639483571053,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 2.962998151779175,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.031317099928855896,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 2.9706358909606934,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.031614501029253006,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 2.9326977729797363,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.031002728268504143,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 2.99625301361084,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.03216253221035004,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 2.920912027359009,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.033380527049303055,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 2.9688234329223633,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.03085758350789547,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 2.9811222553253174,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.03258587792515755,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 2.932366371154785,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.031639691442251205,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 2.988865613937378,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.03144839406013489,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 2.9750723838806152,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.030864672735333443,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 2.999329090118408,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.030679022893309593,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 2.974264144897461,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.03236627206206322,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 2.9808404445648193,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.03115275874733925,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 3.0216290950775146,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.029868317767977715,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 2.9739654064178467,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.03446527197957039,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 2.9565515518188477,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.033536117523908615,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 2.958047389984131,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.03113192319869995,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 2.9662246704101562,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.03089931420981884,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 2.9972119331359863,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.030822310596704483,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 3.020500421524048,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.032264843583106995,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 2.9895882606506348,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.031226349994540215,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 2.958217144012451,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.03110436722636223,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 2.9884791374206543,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.03358910232782364,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 2.9908103942871094,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.03348501771688461,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 2.985677480697632,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.029595404863357544,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 2.979318380355835,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.031459368765354156,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 2.9921798706054688,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.03067564032971859,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 2.977778434753418,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.030506785959005356,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 2.9841768741607666,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.029978664591908455,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 2.995443344116211,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.030900927260518074,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 3.0177016258239746,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.029710259288549423,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 3.01532244682312,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.031190430745482445,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 2.9791834354400635,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.03150483965873718,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 2.998800754547119,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.030249621719121933,
+ "learning_rate": 0.000125422220031917,
+ "loss": 3.0187838077545166,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.031052686274051666,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 2.984062433242798,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.02961024083197117,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 2.970818042755127,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.03020382858812809,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 2.9913196563720703,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.031914591789245605,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 2.993145704269409,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.03061537817120552,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 3.0131571292877197,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.029960963875055313,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 3.0017619132995605,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.030348235741257668,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 3.0220556259155273,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.03030729480087757,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 3.0499038696289062,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.030809229239821434,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 2.986651659011841,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.029438305646181107,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 3.0351314544677734,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.029887216165661812,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 3.021660327911377,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.02955588512122631,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 2.9990694522857666,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.029586469754576683,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 3.0296952724456787,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.030632557347416878,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 3.0124430656433105,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.030579863116145134,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 3.0381762981414795,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.03282996267080307,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 3.0329041481018066,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.031176412478089333,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 3.037598133087158,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.028640294447541237,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 3.0004701614379883,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.0326533280313015,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 3.0328853130340576,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.03041139990091324,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 3.0149893760681152,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.030366195365786552,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 3.0379695892333984,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.03106893040239811,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 3.0284576416015625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.0288887657225132,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 3.028569221496582,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.03154266998171806,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 3.0737850666046143,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.0305064395070076,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 3.0744194984436035,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.030967431142926216,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 3.0170607566833496,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.036945272237062454,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 3.0355610847473145,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.329805374145508,
+ "eval_runtime": 64.1339,
+ "eval_samples_per_second": 38.077,
+ "eval_steps_per_second": 1.201,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.03236053138971329,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 2.9230728149414062,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.0395631417632103,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 2.92578125,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.0361955352127552,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 2.937826633453369,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.035918429493904114,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 2.9481167793273926,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.03309318795800209,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 2.8919506072998047,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.0327315591275692,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 2.903585433959961,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.033141329884529114,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 2.942802667617798,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.033966805785894394,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 2.9210848808288574,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.03218021243810654,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 2.933588981628418,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.03272828459739685,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 2.9202046394348145,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.030764421448111534,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 2.916121006011963,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.031897611916065216,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 2.957679033279419,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.03199796751141548,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 2.9276621341705322,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.030233927071094513,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 2.9006078243255615,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.03035062551498413,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 2.920341730117798,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.03029404580593109,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 2.937875747680664,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.029371997341513634,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 2.9832730293273926,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.030163884162902832,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 2.94118332862854,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.03072655200958252,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 2.9632580280303955,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.02975962497293949,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 2.9346470832824707,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.029814405366778374,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 2.9451403617858887,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.030135225504636765,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 2.944918155670166,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.029395751655101776,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 2.925586223602295,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.0306396484375,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 2.9487359523773193,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.03110465221107006,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 2.963491678237915,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.03155078738927841,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 2.9355576038360596,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.030103784054517746,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 2.953045606613159,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.030827803537249565,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 2.968574047088623,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.0310911163687706,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 2.977902412414551,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.030260782688856125,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 2.94873309135437,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.030654428526759148,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 2.9282753467559814,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.02993479184806347,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 2.97076416015625,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.030244970694184303,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 2.960441827774048,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.030003050342202187,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 2.9672164916992188,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.030057862401008606,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 2.961918830871582,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.029379919171333313,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 2.9918394088745117,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.03060895763337612,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 2.94891619682312,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.030228130519390106,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 2.9726617336273193,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.029833678156137466,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 2.9494335651397705,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.030842263251543045,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 2.9580483436584473,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.029707176610827446,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 2.963684320449829,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.030966786667704582,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 2.9710958003997803,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.03137032687664032,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 3.003589391708374,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.03123190440237522,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 2.9742860794067383,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.030608588829636574,
+ "learning_rate": 0.000110418175419276,
+ "loss": 2.97404408454895,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.030018644407391548,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 2.972231864929199,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.03056846372783184,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 2.9871716499328613,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.02888435684144497,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 2.9774169921875,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.03173990175127983,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 2.941751480102539,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.030047327280044556,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 2.990516185760498,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.030729815363883972,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 2.9745054244995117,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.030600983649492264,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 2.9676642417907715,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.03208782523870468,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 3.006941556930542,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.031162556260824203,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 2.9740395545959473,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.031870320439338684,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 2.9809296131134033,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.03184300288558006,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 2.9781875610351562,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.029629364609718323,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 2.9912891387939453,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.03322628512978554,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 2.9957847595214844,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.0322413295507431,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 2.9894957542419434,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.030753307044506073,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 3.0320589542388916,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.03447900339961052,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 3.0189106464385986,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.032016605138778687,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 3.0162312984466553,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.030838409438729286,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 3.000263214111328,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.034556787461042404,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 2.996680736541748,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.03266173601150513,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 2.990510940551758,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.030973197892308235,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 2.9767255783081055,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.03442178666591644,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 2.9941515922546387,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.03215927258133888,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 3.006618022918701,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.030949831008911133,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 3.004136085510254,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.03256693854928017,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 3.0229580402374268,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.02986278384923935,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 2.9580869674682617,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.03597614914178848,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 2.995757579803467,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.359405517578125,
+ "eval_runtime": 68.1721,
+ "eval_samples_per_second": 35.821,
+ "eval_steps_per_second": 1.129,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.03432514891028404,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 2.922243118286133,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.0345693901181221,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 2.9114043712615967,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0349399633705616,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 2.8668057918548584,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.03161955252289772,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 2.8996763229370117,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.032124556601047516,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 2.8662075996398926,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.03291172534227371,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 2.9088220596313477,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.033552397042512894,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 2.9076967239379883,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.03218761831521988,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 2.8726117610931396,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.03289451450109482,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 2.9127345085144043,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.033019524067640305,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 2.920897960662842,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.03153533488512039,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 2.9124374389648438,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.032293178141117096,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 2.877277374267578,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.031969670206308365,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 2.921635150909424,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.03023488074541092,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 2.9336061477661133,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.03215106576681137,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 2.905877113342285,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.030415233224630356,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 2.909331798553467,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.031150560826063156,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 2.9302995204925537,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.03152180463075638,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 2.9171547889709473,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.02996763028204441,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 2.916459798812866,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.030544105917215347,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 2.898505449295044,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.030963370576500893,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 2.948704719543457,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.03191790357232094,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 2.9182960987091064,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.02956460975110531,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 2.9211082458496094,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.031227655708789825,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 2.942269802093506,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.031859587877988815,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 2.9169349670410156,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.03045276179909706,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 2.9240684509277344,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.031577471643686295,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 2.9135894775390625,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.031031791120767593,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 2.913355827331543,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.030012110248208046,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 2.8860886096954346,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.029772799462080002,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 2.974332809448242,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.029961293563246727,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 2.943694591522217,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.029537590220570564,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 2.93180513381958,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.030834559351205826,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 2.9442358016967773,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.02998523600399494,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 2.9336910247802734,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.030030962079763412,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 2.93593692779541,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.030250679701566696,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 2.9370553493499756,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.02942339889705181,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 2.9406495094299316,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.030448907986283302,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 2.9552698135375977,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.03078317642211914,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 2.933631420135498,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.029188675805926323,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 2.944002628326416,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.03153637796640396,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 2.9708166122436523,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.030299564823508263,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 2.9570717811584473,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.03169458732008934,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 2.942774772644043,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.031527288258075714,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 2.9711742401123047,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.0314360186457634,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 2.9509239196777344,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.032102152705192566,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 2.9443840980529785,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.03111223131418228,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 2.972594738006592,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.0322272814810276,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 2.9361982345581055,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.033640168607234955,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 2.9726009368896484,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.02979348786175251,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 2.9978041648864746,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.030464721843600273,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 2.949810028076172,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.03170578181743622,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 2.9677181243896484,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.02996704913675785,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 2.974087953567505,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.030277637764811516,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 2.95644211769104,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.030859284102916718,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 2.957258701324463,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.029737796634435654,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 2.9502854347229004,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.029822375625371933,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 2.9625749588012695,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.030455606058239937,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 2.9669153690338135,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.029491793364286423,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 2.952991247177124,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.030326122418045998,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 2.978355646133423,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.03011198714375496,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 2.976996421813965,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.029053322970867157,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 2.9754371643066406,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.02925761416554451,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 2.94271183013916,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.029678862541913986,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 2.9386143684387207,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.031623054295778275,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 2.9660329818725586,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.030133802443742752,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 2.9580249786376953,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.029482144862413406,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 2.944474697113037,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.030708568170666695,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 2.967862367630005,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.028793595731258392,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 2.9581642150878906,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.03083081543445587,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 2.957261323928833,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.030180038884282112,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 2.9579484462738037,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.034881673753261566,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 2.9007070064544678,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.38953161239624,
+ "eval_runtime": 71.4964,
+ "eval_samples_per_second": 34.156,
+ "eval_steps_per_second": 1.077,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0335528701543808,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 2.8542537689208984,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.03163984417915344,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 2.864593505859375,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.034005917608737946,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 2.855072498321533,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.03235425800085068,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 2.8504371643066406,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.03171464428305626,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 2.8924291133880615,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.0333196297287941,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 2.8764119148254395,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.03050111047923565,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 2.874119281768799,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.032114267349243164,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 2.9059693813323975,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.03141588345170021,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 2.879429340362549,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.03201865404844284,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 2.898519277572632,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.03152940049767494,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 2.866131544113159,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.032520636916160583,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 2.888209819793701,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.03244909271597862,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 2.888913631439209,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.03160687908530235,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 2.905120372772217,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.031518299132585526,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 2.8985815048217773,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.0304370429366827,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 2.8531291484832764,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.03239843621850014,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 2.876098394393921,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.03194732964038849,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 2.908271074295044,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.03145617991685867,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 2.874128580093384,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.031182898208498955,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 2.894904613494873,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.0323818176984787,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 2.9005861282348633,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.030969973653554916,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 2.9236338138580322,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.03238993510603905,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 2.8811020851135254,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.030876649543642998,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 2.8802058696746826,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.032251305878162384,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 2.882112503051758,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.03334304690361023,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 2.9290661811828613,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.03299225866794586,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 2.908268928527832,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.03335456922650337,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 2.9156572818756104,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.030786098912358284,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 2.8981709480285645,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.03379470854997635,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 2.9483864307403564,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.03020016849040985,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 2.908958911895752,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.032774876803159714,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 2.899505138397217,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.030589981004595757,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 2.9257779121398926,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.03366618603467941,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 2.8749284744262695,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.03038695827126503,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 2.9127252101898193,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.03400666266679764,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 2.922375202178955,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.031702589243650436,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 2.8886446952819824,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.03223840892314911,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 2.906278610229492,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.0327419750392437,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 2.9436042308807373,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.030698910355567932,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 2.897615909576416,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.03250030800700188,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 2.928028106689453,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.030623450875282288,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 2.9268975257873535,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.030981825664639473,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 2.938185691833496,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.03139727562665939,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 2.9397618770599365,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.029906271025538445,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 2.921111583709717,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.03110392577946186,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 2.899641513824463,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.03060143068432808,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 2.922635078430176,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.03061371110379696,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 2.925189733505249,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.03080526366829872,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 2.90329647064209,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.030412547290325165,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 2.9444844722747803,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.03024943359196186,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 2.9165825843811035,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.030653616413474083,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 2.9171371459960938,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.03027702495455742,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 2.9399702548980713,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.030466679483652115,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 2.888066291809082,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.02997267246246338,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 2.906545877456665,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.03026283159852028,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 2.9359335899353027,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.0293166171759367,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 2.9118804931640625,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.031117571517825127,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 2.968198299407959,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.03041299246251583,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 2.9333484172821045,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.03037327155470848,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 2.895859956741333,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.03161945194005966,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 2.9491944313049316,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.031058384105563164,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 2.926038980484009,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.030037537217140198,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 2.941394805908203,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.030748063698410988,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 2.998904228210449,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.030324939638376236,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 2.9226560592651367,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.030745431780815125,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 2.977520227432251,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.02952338382601738,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 2.9365828037261963,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.030132673680782318,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 2.9574036598205566,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.030379803851246834,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 2.9329113960266113,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.029670827090740204,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 2.951117515563965,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.029823293909430504,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 2.9586987495422363,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.03546741604804993,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 2.9378662109375,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.417913436889648,
+ "eval_runtime": 69.6157,
+ "eval_samples_per_second": 35.078,
+ "eval_steps_per_second": 1.106,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.033859945833683014,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 2.8493661880493164,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.032931916415691376,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 2.844602108001709,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.03232948109507561,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 2.8463873863220215,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.032815027981996536,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 2.846658706665039,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.03269147500395775,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 2.8286149501800537,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.03337012976408005,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 2.9202351570129395,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.031733594834804535,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 2.8601999282836914,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.032877493649721146,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 2.8566389083862305,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.033852867782115936,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 2.859797477722168,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.03143248334527016,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 2.8411474227905273,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.033198583871126175,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 2.888702392578125,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.029946891590952873,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 2.8332901000976562,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.031660862267017365,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 2.8255481719970703,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.030189910903573036,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 2.8664190769195557,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.03273668885231018,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 2.886531114578247,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.030584124848246574,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 2.872990131378174,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.032543931156396866,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 2.891528606414795,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.031016048043966293,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 2.8759872913360596,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.03190434351563454,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 2.8757526874542236,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.030996622517704964,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 2.862654447555542,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.03289523348212242,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 2.9064857959747314,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.03322640061378479,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 2.8921048641204834,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.0303682629019022,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 2.886458396911621,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.033412594348192215,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 2.8638253211975098,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.030457086861133575,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 2.835578441619873,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.03225106745958328,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 2.8652632236480713,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.03178466483950615,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 2.8796353340148926,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.03075973503291607,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 2.895150661468506,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.02984696812927723,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 2.882878303527832,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.030945835635066032,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 2.898838520050049,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.030904723331332207,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 2.885678291320801,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.030709490180015564,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 2.8645198345184326,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.030528701841831207,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 2.8877408504486084,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.030471473932266235,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 2.9136905670166016,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.030309896916151047,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 2.9238810539245605,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.030605493113398552,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 2.8624460697174072,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.029686832800507545,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 2.9085543155670166,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.030039243400096893,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 2.8624000549316406,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.02953776717185974,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 2.8893938064575195,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.029651308432221413,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 2.914975166320801,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.030902806669473648,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 2.887071132659912,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.02948773466050625,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 2.8893041610717773,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.03049607202410698,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 2.902736186981201,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.030077926814556122,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 2.8936967849731445,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.0303986556828022,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 2.9223172664642334,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.030650658532977104,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 2.9140920639038086,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.029562659561634064,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 2.918179988861084,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.03063463792204857,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 2.9222826957702637,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.030284354463219643,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 2.8685007095336914,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.029791876673698425,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 2.914102554321289,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.030152756720781326,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 2.8890669345855713,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.02957645244896412,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 2.8851962089538574,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.02963421493768692,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 2.9012060165405273,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.029279837384819984,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 2.949984073638916,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.0302849430590868,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 2.9049415588378906,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.029432371258735657,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 2.8919944763183594,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.030542058870196342,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 2.9168801307678223,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.02911970019340515,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 2.9342129230499268,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.030335791409015656,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 2.918611764907837,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.03031272068619728,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 2.8769750595092773,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.030200788751244545,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 2.8994460105895996,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.03032573312520981,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 2.920720100402832,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.03180326521396637,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 2.891145706176758,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.029451677575707436,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 2.9057083129882812,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.030261319130659103,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 2.9132461547851562,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.031257566064596176,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 2.9089136123657227,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.03043205663561821,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 2.912537097930908,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.030708536505699158,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 2.8841137886047363,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.03044111281633377,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 2.921257257461548,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.029576802626252174,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 2.908468723297119,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.03031427040696144,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 2.936445713043213,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.036312974989414215,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 2.8997855186462402,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.443755149841309,
+ "eval_runtime": 62.9898,
+ "eval_samples_per_second": 38.768,
+ "eval_steps_per_second": 1.222,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.03337511047720909,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 2.8584513664245605,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.031415343284606934,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 2.8529202938079834,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.033063437789678574,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 2.814058780670166,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.03153088316321373,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 2.80460786819458,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.03227711841464043,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 2.858191967010498,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.030423127114772797,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 2.845815420150757,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.032030411064624786,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 2.837918281555176,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.030464086681604385,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 2.841099739074707,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.03194798156619072,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 2.843568801879883,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.03152863308787346,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 2.8779497146606445,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.03201183304190636,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 2.906585216522217,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.02999979443848133,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 2.84859561920166,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.03106156550347805,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 2.814643383026123,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.03118651732802391,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 2.8370001316070557,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.031022636219859123,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 2.8466384410858154,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.03163835406303406,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 2.808440685272217,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.03071477636694908,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 2.850621223449707,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.03041849471628666,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 2.818002939224243,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.03104640357196331,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 2.8375182151794434,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.02970169298350811,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 2.871188163757324,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.03146181255578995,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 2.8356661796569824,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.030204759910702705,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 2.861675262451172,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.029957789927721024,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 2.8493146896362305,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.031224241480231285,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 2.876760959625244,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 0.02914278209209442,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 2.838275909423828,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.03204638510942459,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 2.8679757118225098,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.030257491394877434,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 2.8245105743408203,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.031307291239500046,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 2.852024793624878,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 0.030594198033213615,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 2.883610248565674,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.03119277022778988,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 2.866126775741577,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.0302574560046196,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 2.857886791229248,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.029702702537178993,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 2.8324077129364014,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.029572894796729088,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 2.878951072692871,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.03024783544242382,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 2.900284767150879,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.02952091582119465,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 2.8402292728424072,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.03006756864488125,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 2.8609299659729004,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.029146816581487656,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 2.887960433959961,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.03026733547449112,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 2.852726459503174,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.029839467257261276,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 2.8559489250183105,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.029919864609837532,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 2.880255937576294,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.03074759431183338,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 2.8586673736572266,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.029141079634428024,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 2.8813042640686035,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.031023059040308,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 2.8355565071105957,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.029786985367536545,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 2.872175693511963,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.030190780758857727,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 2.874236583709717,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.029669199138879776,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 2.8723950386047363,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.030073825269937515,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 2.863720417022705,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.02976175583899021,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 2.883267402648926,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.030071711167693138,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 2.8690664768218994,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.031068995594978333,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 2.9107182025909424,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.030393972992897034,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 2.859959602355957,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.030435480177402496,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 2.9003896713256836,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.02982497587800026,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 2.857231378555298,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.030999863520264626,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 2.8567168712615967,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.029698362573981285,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 2.8949484825134277,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.031584806740283966,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 2.8703250885009766,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.02952221967279911,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 2.8914244174957275,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.03138599172234535,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 2.917753219604492,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.029678434133529663,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 2.8854236602783203,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.03174183890223503,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 2.8508100509643555,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.03006329946219921,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 2.883819103240967,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.031188689172267914,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 2.932220458984375,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.029729023575782776,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 2.894927978515625,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.02953082136809826,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 2.896371364593506,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.030423447489738464,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 2.9003541469573975,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.029901325702667236,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 2.8726208209991455,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.030071355402469635,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 2.881291151046753,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.029847780242562294,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 2.944478988647461,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.029794953763484955,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 2.8625364303588867,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.03029358573257923,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 2.9106764793395996,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.031095417216420174,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 2.8879761695861816,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.03671329468488693,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 2.877009391784668,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 4.470407962799072,
+ "eval_runtime": 57.8161,
+ "eval_samples_per_second": 42.237,
+ "eval_steps_per_second": 1.332,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.03383656591176987,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 2.8301892280578613,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.030587317422032356,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 2.809070587158203,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.034895412623882294,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 2.845365285873413,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.03129304572939873,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 2.8176681995391846,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.03369338437914848,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 2.820741653442383,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.030895931646227837,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 2.810272693634033,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.03239824250340462,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 2.855811595916748,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.03123638592660427,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 2.7941160202026367,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.031117001548409462,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 2.784010887145996,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.03162471950054169,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 2.8337666988372803,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.0316263884305954,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 2.8019790649414062,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.031872715801000595,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 2.8156003952026367,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.030252376571297646,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 2.7971153259277344,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.03168163076043129,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 2.791532039642334,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.03014831617474556,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 2.8359432220458984,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.03142879158258438,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 2.846064567565918,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.030883461236953735,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 2.8668103218078613,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.03189866617321968,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 2.8214638233184814,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.029116913676261902,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 2.850940704345703,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.030518673360347748,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 2.844576358795166,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.029227688908576965,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 2.797168493270874,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.029839880764484406,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 2.8534903526306152,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.03011847473680973,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 2.8278157711029053,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.03027244098484516,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 2.858158588409424,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.030171507969498634,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 2.8523292541503906,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.029657812789082527,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 2.8015360832214355,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.028928041458129883,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 2.8357324600219727,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.029674772173166275,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 2.8269171714782715,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.03013371117413044,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 2.822054862976074,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.02960435301065445,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 2.8452866077423096,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.030388740822672844,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 2.855799674987793,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.029823970049619675,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 2.8262312412261963,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.030613111332058907,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 2.8490824699401855,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.029467277228832245,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 2.855149984359741,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.030264565721154213,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 2.849930763244629,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.02961486577987671,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 2.843088150024414,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.029051944613456726,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 2.825998067855835,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.03038879856467247,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 2.8565316200256348,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.02917695790529251,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 2.8733534812927246,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.03011898323893547,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 2.8347158432006836,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.03015534020960331,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 2.878455400466919,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.02992299199104309,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 2.861504554748535,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.030217986553907394,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 2.8451390266418457,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.029564134776592255,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 2.8065595626831055,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.03130771592259407,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 2.8506879806518555,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.030411556363105774,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 2.857304573059082,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.030868662521243095,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 2.8545470237731934,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.03002075105905533,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 2.889376640319824,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.030850842595100403,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 2.8634769916534424,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.03078618459403515,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 2.8487653732299805,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.031029092147946358,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 2.897327184677124,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.03101371042430401,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 2.8642020225524902,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.03044370748102665,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 2.87313175201416,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.030529852956533432,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 2.849147081375122,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.029675021767616272,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 2.8475494384765625,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.03002132847905159,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 2.8713879585266113,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.03048008307814598,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 2.8976492881774902,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.02930760383605957,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 2.850775718688965,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.029840776696801186,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 2.8743128776550293,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.030162667855620384,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 2.850478410720825,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.029736310243606567,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 2.851759195327759,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.029385298490524292,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 2.8473241329193115,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.028572317212820053,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 2.852155923843384,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.0300681721419096,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 2.871021270751953,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.0292969960719347,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 2.8470516204833984,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.029493169859051704,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 2.892655372619629,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.029351307079195976,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 2.8886585235595703,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.029827935621142387,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 2.8802051544189453,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.029231170192360878,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 2.876192092895508,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.029604196548461914,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 2.8420169353485107,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.030229991301894188,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 2.86871075630188,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.0357736200094223,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 2.852806568145752,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 4.496274948120117,
+ "eval_runtime": 58.0976,
+ "eval_samples_per_second": 42.033,
+ "eval_steps_per_second": 1.325,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.03180048242211342,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 2.85683012008667,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.030023347586393356,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 2.814507484436035,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.03346574679017067,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 2.8209877014160156,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.030521582812070847,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 2.7859511375427246,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.033295996487140656,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 2.7879018783569336,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.029781939461827278,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 2.801133155822754,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.033207762986421585,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 2.8046321868896484,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.03027796931564808,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 2.8235392570495605,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.03096267767250538,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 2.8181068897247314,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.030886022374033928,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 2.7954399585723877,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.030112111940979958,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 2.78225040435791,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.03067166730761528,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 2.8147826194763184,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.029821906238794327,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 2.811027765274048,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.030433299019932747,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 2.8178303241729736,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.02960469201207161,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 2.802370548248291,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.029496584087610245,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 2.8156142234802246,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.029699044302105904,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 2.7911806106567383,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.030225902795791626,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 2.7872819900512695,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.02939198724925518,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 2.800001621246338,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.03059281036257744,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 2.8127222061157227,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.02927992306649685,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 2.8320517539978027,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.03038649819791317,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 2.8169116973876953,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.029806654900312424,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 2.8279008865356445,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.029620900750160217,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 2.8212599754333496,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.029498178511857986,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 2.8112261295318604,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.02918926067650318,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 2.8290767669677734,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.029827842488884926,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 2.8396403789520264,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.029516329988837242,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 2.840583324432373,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.029057780280709267,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 2.835113525390625,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.029345065355300903,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 2.8197970390319824,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.02873479388654232,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 2.8052587509155273,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.029737083241343498,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 2.819146156311035,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.0292612686753273,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 2.8543701171875,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.028945421800017357,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 2.804084300994873,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.029137294739484787,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 2.7658708095550537,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.02911440096795559,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 2.8134429454803467,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.028932519257068634,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 2.810455322265625,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.029404157772660255,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 2.802886962890625,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.028970446437597275,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 2.8426849842071533,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.029187830165028572,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 2.8366644382476807,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.029293285682797432,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 2.8261566162109375,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.029366638511419296,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 2.8125827312469482,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.02900587022304535,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 2.839684009552002,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.029162462800741196,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 2.8331997394561768,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.029060618951916695,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 2.8627192974090576,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.02887798845767975,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 2.859687089920044,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.02894580364227295,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 2.8195912837982178,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.029554955661296844,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 2.852871894836426,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.02822914719581604,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 2.8671388626098633,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.02954651229083538,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 2.8524365425109863,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.02902315929532051,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 2.843954563140869,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.0290356632322073,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 2.8504233360290527,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.029262971132993698,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 2.8657913208007812,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.029469065368175507,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 2.82297945022583,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.029609832912683487,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 2.856297016143799,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.029690174385905266,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 2.8302533626556396,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.03018389455974102,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 2.8542540073394775,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.029053693637251854,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 2.8432576656341553,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.028549278154969215,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 2.8324697017669678,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.030014613643288612,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 2.800536870956421,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.02949996292591095,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 2.835141897201538,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.028825247660279274,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 2.8263447284698486,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.029436098411679268,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 2.8522257804870605,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.029510991647839546,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 2.851526975631714,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.028975380584597588,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 2.820462465286255,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.029128702357411385,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 2.861868381500244,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.02951551228761673,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 2.875210762023926,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.028196118772029877,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 2.8345048427581787,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.029215430840849876,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 2.84328556060791,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.02857460267841816,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 2.8448305130004883,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.02842608466744423,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 2.860792636871338,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03630338981747627,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 2.8723089694976807,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 4.522201061248779,
+ "eval_runtime": 57.3764,
+ "eval_samples_per_second": 42.561,
+ "eval_steps_per_second": 1.342,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.030005959793925285,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 2.7629683017730713,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.028917059302330017,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 2.7894129753112793,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.029286719858646393,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 2.8119699954986572,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.02967856638133526,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 2.8115768432617188,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.029460743069648743,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 2.7809553146362305,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.028186975046992302,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 2.818805456161499,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.02975514344871044,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 2.816215991973877,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.029532121494412422,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 2.782461166381836,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.02879161946475506,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 2.768501043319702,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.029776159673929214,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 2.7995424270629883,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.028667304664850235,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 2.773409843444824,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.028793245553970337,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 2.7938318252563477,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.029090600088238716,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 2.823000907897949,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.028883792459964752,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 2.8175344467163086,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.028518877923488617,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 2.792815685272217,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.029769321903586388,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 2.831251382827759,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.029009172692894936,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 2.8345084190368652,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.028994156047701836,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 2.7925689220428467,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.029505716636776924,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 2.814758777618408,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.028683487325906754,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 2.7954044342041016,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.029633600264787674,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 2.794569969177246,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.028943300247192383,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 2.803800106048584,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.029026521369814873,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 2.760383129119873,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.029057513922452927,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 2.8081226348876953,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.02906658686697483,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 2.8183770179748535,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.02926407754421234,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 2.831427574157715,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.029171453788876534,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 2.826587200164795,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.029656723141670227,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 2.7633590698242188,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.029490994289517403,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 2.7932424545288086,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.0291811041533947,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 2.8346633911132812,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.029164446517825127,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 2.8193140029907227,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.029394259676337242,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 2.8272342681884766,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.029221080243587494,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 2.79146146774292,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.028165455907583237,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 2.7826766967773438,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.0296048391610384,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 2.8119699954986572,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.028316112235188484,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 2.824860095977783,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.02878737449645996,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 2.811997890472412,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.029061628505587578,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 2.8370048999786377,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.028607595711946487,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 2.8093791007995605,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.028144560754299164,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 2.8236052989959717,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.028813332319259644,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 2.838690996170044,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.028437303379178047,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 2.800854206085205,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.028567243367433548,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 2.8301069736480713,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.029187796637415886,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 2.823208808898926,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.028203707188367844,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 2.804643154144287,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.02822236344218254,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 2.8132784366607666,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.028921568766236305,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 2.8215417861938477,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.02867986261844635,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 2.8318984508514404,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.028849760070443153,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 2.8209228515625,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.028631621971726418,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 2.8219287395477295,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.028188595548272133,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 2.814027786254883,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.028504883870482445,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 2.847404956817627,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.02826615609228611,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 2.808562755584717,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.02844209223985672,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 2.827244520187378,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.028765909373760223,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 2.81117844581604,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.028130069375038147,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 2.852351188659668,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.028408419340848923,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 2.847090244293213,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.028523078188300133,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 2.821716785430908,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.028461646288633347,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 2.816019296646118,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.028531840071082115,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 2.8226165771484375,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.028871113434433937,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 2.831231117248535,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.0285959430038929,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 2.8257393836975098,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.02841980569064617,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 2.8298444747924805,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.028909973800182343,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 2.808438777923584,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.02884981967508793,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 2.815157890319824,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.02836848609149456,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 2.792750835418701,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.028464308008551598,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 2.7594008445739746,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.02909877710044384,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 2.844888925552368,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.0285524632781744,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 2.8162052631378174,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.02916364185512066,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 2.8175275325775146,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.028360974043607712,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 2.8322806358337402,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.0366419292986393,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 2.8780605792999268,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 4.539337635040283,
+ "eval_runtime": 57.482,
+ "eval_samples_per_second": 42.483,
+ "eval_steps_per_second": 1.34,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.030007537454366684,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 2.7774059772491455,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.028395716100931168,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 2.8240408897399902,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.029259607195854187,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 2.7780704498291016,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.029434163123369217,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 2.802172899246216,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.02914051152765751,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 2.746731758117676,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.02971787191927433,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 2.7998170852661133,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.029050175100564957,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 2.786863327026367,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.030421903356909752,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 2.789545774459839,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.029742848128080368,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 2.827561855316162,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.03001682087779045,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 2.8226547241210938,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.030296841636300087,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 2.795665740966797,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.02908152900636196,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 2.7598729133605957,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.029059268534183502,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 2.786234140396118,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.02878507599234581,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 2.7754065990448,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.02881552092730999,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 2.779616355895996,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.02847588248550892,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 2.7556540966033936,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.028446536511182785,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 2.7858777046203613,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.028488771989941597,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 2.8308897018432617,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.028518928214907646,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 2.798438310623169,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.02807544358074665,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 2.7854204177856445,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.028449799865484238,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 2.787806510925293,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.02806367352604866,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 2.773944139480591,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.028127942234277725,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 2.795468807220459,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.028206733986735344,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 2.794552803039551,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.02769204042851925,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 2.78456974029541,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.028235256671905518,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 2.8258988857269287,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.028410933911800385,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 2.8145394325256348,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.028073569759726524,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 2.739804744720459,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.027985481545329094,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 2.7385730743408203,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.0285568255931139,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 2.790156364440918,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.028381718322634697,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 2.7715389728546143,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.028455479070544243,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 2.804957628250122,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.028330320492386818,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 2.813432216644287,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.02832239493727684,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 2.8087422847747803,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.028423218056559563,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 2.790351390838623,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.028114303946495056,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 2.8367419242858887,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.028525900095701218,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 2.8024988174438477,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.027810173109173775,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 2.8156440258026123,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.028172023594379425,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 2.801828384399414,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.027999311685562134,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 2.7917897701263428,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.02863648533821106,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 2.7998974323272705,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.027806704863905907,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 2.7955031394958496,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.028083806857466698,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 2.8160462379455566,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.02817711979150772,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 2.7673726081848145,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.027995774522423744,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 2.787623882293701,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.028293436393141747,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 2.848489284515381,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.028710173442959785,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 2.8186140060424805,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.028271062299609184,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 2.822369337081909,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.02830388769507408,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 2.8050756454467773,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.02831537462770939,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 2.8161096572875977,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.028237098827958107,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 2.7897520065307617,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.027944844216108322,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 2.816229820251465,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.028575003147125244,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 2.8069968223571777,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.028569547459483147,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 2.78981614112854,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.028383128345012665,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 2.820969343185425,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.027872422710061073,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 2.781301498413086,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.028342951089143753,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 2.816469192504883,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.028373923152685165,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 2.799290657043457,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.02807556465268135,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 2.814924716949463,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.02789521962404251,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 2.7950439453125,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.028297580778598785,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 2.8535611629486084,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.02745051123201847,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 2.7812037467956543,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.027246123179793358,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 2.778583288192749,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.027801746502518654,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 2.812328815460205,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.027761677280068398,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 2.829758882522583,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.028340548276901245,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 2.8203043937683105,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.027355236932635307,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 2.777925968170166,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.027668653056025505,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 2.789498805999756,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.02823868952691555,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 2.7765135765075684,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.02812831476330757,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 2.774756669998169,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.027789659798145294,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 2.8317346572875977,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.035616710782051086,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 2.8123791217803955,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 4.558599948883057,
+ "eval_runtime": 56.3211,
+ "eval_samples_per_second": 43.359,
+ "eval_steps_per_second": 1.367,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.027842728421092033,
+ "learning_rate": 2.1371380147693752e-05,
+ "loss": 2.766806125640869,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.027614781633019447,
+ "learning_rate": 2.1286841417243662e-05,
+ "loss": 2.7721548080444336,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.027729114517569542,
+ "learning_rate": 2.1202529657997137e-05,
+ "loss": 2.775094509124756,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.027813846245408058,
+ "learning_rate": 2.1118444937882538e-05,
+ "loss": 2.7677454948425293,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.027814052999019623,
+ "learning_rate": 2.1034587324645456e-05,
+ "loss": 2.764620304107666,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.027886489406228065,
+ "learning_rate": 2.095095688584846e-05,
+ "loss": 2.736621141433716,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.027611345052719116,
+ "learning_rate": 2.086755368887107e-05,
+ "loss": 2.775693893432617,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.027603253722190857,
+ "learning_rate": 2.0784377800909802e-05,
+ "loss": 2.7538721561431885,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.027774354442954063,
+ "learning_rate": 2.0701429288977847e-05,
+ "loss": 2.7577450275421143,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.028701378032565117,
+ "learning_rate": 2.0618708219905455e-05,
+ "loss": 2.792163848876953,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.02745729312300682,
+ "learning_rate": 2.0536214660339398e-05,
+ "loss": 2.7608468532562256,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.02753688395023346,
+ "learning_rate": 2.0453948676743267e-05,
+ "loss": 2.7727649211883545,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.027698034420609474,
+ "learning_rate": 2.037191033539731e-05,
+ "loss": 2.771090269088745,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.028323858976364136,
+ "learning_rate": 2.029009970239831e-05,
+ "loss": 2.7989578247070312,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.027763206511735916,
+ "learning_rate": 2.0208516843659678e-05,
+ "loss": 2.8134682178497314,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.027521144598722458,
+ "learning_rate": 2.012716182491114e-05,
+ "loss": 2.827221393585205,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.02790055237710476,
+ "learning_rate": 2.0046034711699048e-05,
+ "loss": 2.7750606536865234,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.027743415907025337,
+ "learning_rate": 1.9965135569386032e-05,
+ "loss": 2.754495620727539,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.02780855819582939,
+ "learning_rate": 1.9884464463151077e-05,
+ "loss": 2.793025255203247,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.027743151411414146,
+ "learning_rate": 1.980402145798933e-05,
+ "loss": 2.7627084255218506,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.02772294543683529,
+ "learning_rate": 1.9723806618712484e-05,
+ "loss": 2.810464382171631,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.027665726840496063,
+ "learning_rate": 1.9643820009947983e-05,
+ "loss": 2.7772934436798096,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.027962513267993927,
+ "learning_rate": 1.9564061696139687e-05,
+ "loss": 2.791640520095825,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.028066501021385193,
+ "learning_rate": 1.948453174154744e-05,
+ "loss": 2.7920327186584473,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.027441052719950676,
+ "learning_rate": 1.9405230210247042e-05,
+ "loss": 2.797529458999634,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.02827683836221695,
+ "learning_rate": 1.9326157166130365e-05,
+ "loss": 2.8248825073242188,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.027658650651574135,
+ "learning_rate": 1.9247312672905017e-05,
+ "loss": 2.8131494522094727,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.0280227642506361,
+ "learning_rate": 1.9168696794094758e-05,
+ "loss": 2.7728159427642822,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.02791561745107174,
+ "learning_rate": 1.9090309593038812e-05,
+ "loss": 2.7671456336975098,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.02787322923541069,
+ "learning_rate": 1.9012151132892425e-05,
+ "loss": 2.7740323543548584,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.027309276163578033,
+ "learning_rate": 1.893422147662644e-05,
+ "loss": 2.7773780822753906,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.027728214859962463,
+ "learning_rate": 1.8856520687027325e-05,
+ "loss": 2.7922744750976562,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.027990298345685005,
+ "learning_rate": 1.8779048826697318e-05,
+ "loss": 2.8120408058166504,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.02773648127913475,
+ "learning_rate": 1.8701805958054e-05,
+ "loss": 2.7819745540618896,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.027117537334561348,
+ "learning_rate": 1.8624792143330544e-05,
+ "loss": 2.781357765197754,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.027264263480901718,
+ "learning_rate": 1.8548007444575656e-05,
+ "loss": 2.7825074195861816,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.027638141065835953,
+ "learning_rate": 1.8471451923653322e-05,
+ "loss": 2.809204578399658,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.027160588651895523,
+ "learning_rate": 1.8395125642242987e-05,
+ "loss": 2.784580707550049,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.02765049785375595,
+ "learning_rate": 1.8319028661839343e-05,
+ "loss": 2.7796647548675537,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.027509981766343117,
+ "learning_rate": 1.8243161043752384e-05,
+ "loss": 2.802097797393799,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.02717861346900463,
+ "learning_rate": 1.8167522849107228e-05,
+ "loss": 2.790086269378662,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.027462702244520187,
+ "learning_rate": 1.809211413884418e-05,
+ "loss": 2.779426336288452,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.027430517598986626,
+ "learning_rate": 1.8016934973718762e-05,
+ "loss": 2.7842659950256348,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.027589300647377968,
+ "learning_rate": 1.7941985414301403e-05,
+ "loss": 2.812390089035034,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.02773849479854107,
+ "learning_rate": 1.786726552097766e-05,
+ "loss": 2.7788920402526855,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.027417341247200966,
+ "learning_rate": 1.7792775353947908e-05,
+ "loss": 2.766779899597168,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.027108531445264816,
+ "learning_rate": 1.7718514973227634e-05,
+ "loss": 2.774667501449585,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.02787691541016102,
+ "learning_rate": 1.7644484438647057e-05,
+ "loss": 2.8050103187561035,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.028047434985637665,
+ "learning_rate": 1.75706838098512e-05,
+ "loss": 2.795063018798828,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.027433287352323532,
+ "learning_rate": 1.749711314629992e-05,
+ "loss": 2.802816390991211,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.027421964332461357,
+ "learning_rate": 1.742377250726779e-05,
+ "loss": 2.8223254680633545,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.027546996250748634,
+ "learning_rate": 1.7350661951844065e-05,
+ "loss": 2.780585765838623,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.027431374415755272,
+ "learning_rate": 1.727778153893248e-05,
+ "loss": 2.787783145904541,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.027354050427675247,
+ "learning_rate": 1.720513132725162e-05,
+ "loss": 2.7926011085510254,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.02790139429271221,
+ "learning_rate": 1.713271137533433e-05,
+ "loss": 2.7482681274414062,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.027517175301909447,
+ "learning_rate": 1.7060521741528127e-05,
+ "loss": 2.791224956512451,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.027358947321772575,
+ "learning_rate": 1.698856248399482e-05,
+ "loss": 2.774211883544922,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.027352966368198395,
+ "learning_rate": 1.6916833660710785e-05,
+ "loss": 2.8025760650634766,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.02778322622179985,
+ "learning_rate": 1.6845335329466606e-05,
+ "loss": 2.7977163791656494,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.027194228023290634,
+ "learning_rate": 1.6774067547867113e-05,
+ "loss": 2.7744717597961426,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.027206066995859146,
+ "learning_rate": 1.6703030373331585e-05,
+ "loss": 2.7848451137542725,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.027613313868641853,
+ "learning_rate": 1.6632223863093328e-05,
+ "loss": 2.795206308364868,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.027662355452775955,
+ "learning_rate": 1.656164807419988e-05,
+ "loss": 2.771212577819824,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.027272798120975494,
+ "learning_rate": 1.6491303063512866e-05,
+ "loss": 2.850273609161377,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.027627959847450256,
+ "learning_rate": 1.6421188887708013e-05,
+ "loss": 2.7897393703460693,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.027814874425530434,
+ "learning_rate": 1.6351305603275067e-05,
+ "loss": 2.8047971725463867,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.027428552508354187,
+ "learning_rate": 1.6281653266517658e-05,
+ "loss": 2.7656729221343994,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.02777225710451603,
+ "learning_rate": 1.6212231933553485e-05,
+ "loss": 2.7727203369140625,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.02802879735827446,
+ "learning_rate": 1.6143041660313997e-05,
+ "loss": 2.847011089324951,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.02760482393205166,
+ "learning_rate": 1.6074082502544668e-05,
+ "loss": 2.7733654975891113,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.02776470221579075,
+ "learning_rate": 1.6005354515804503e-05,
+ "loss": 2.7941837310791016,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.03535662591457367,
+ "learning_rate": 1.59368577554664e-05,
+ "loss": 2.787112236022949,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 4.572118282318115,
+ "eval_runtime": 56.9428,
+ "eval_samples_per_second": 42.885,
+ "eval_steps_per_second": 1.352,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.026943478733301163,
+ "learning_rate": 1.5868592276717137e-05,
+ "loss": 2.7454490661621094,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.027496816590428352,
+ "learning_rate": 1.5800558134556806e-05,
+ "loss": 2.783641815185547,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.026721451431512833,
+ "learning_rate": 1.5732755383799315e-05,
+ "loss": 2.7537829875946045,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.027940884232521057,
+ "learning_rate": 1.5665184079072162e-05,
+ "loss": 2.7764663696289062,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.027781952172517776,
+ "learning_rate": 1.5597844274816352e-05,
+ "loss": 2.776278495788574,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.027947992086410522,
+ "learning_rate": 1.553073602528626e-05,
+ "loss": 2.7712326049804688,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.027257733047008514,
+ "learning_rate": 1.546385938454982e-05,
+ "loss": 2.774122714996338,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.026782797649502754,
+ "learning_rate": 1.5397214406488374e-05,
+ "loss": 2.7600014209747314,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.02728961408138275,
+ "learning_rate": 1.533080114479657e-05,
+ "loss": 2.7231063842773438,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.027733273804187775,
+ "learning_rate": 1.5264619652982352e-05,
+ "loss": 2.807209014892578,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.02731870673596859,
+ "learning_rate": 1.5198669984366876e-05,
+ "loss": 2.7994370460510254,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.02707100473344326,
+ "learning_rate": 1.5132952192084793e-05,
+ "loss": 2.762448787689209,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.027301348745822906,
+ "learning_rate": 1.5067466329083533e-05,
+ "loss": 2.754739284515381,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.02716146782040596,
+ "learning_rate": 1.5002212448123997e-05,
+ "loss": 2.7493300437927246,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.02700486034154892,
+ "learning_rate": 1.4937190601780052e-05,
+ "loss": 2.767707347869873,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.0270010344684124,
+ "learning_rate": 1.4872400842438576e-05,
+ "loss": 2.7867441177368164,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.02769402787089348,
+ "learning_rate": 1.4807843222299557e-05,
+ "loss": 2.7755560874938965,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.027279293164610863,
+ "learning_rate": 1.4743517793375814e-05,
+ "loss": 2.7624804973602295,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.0273398756980896,
+ "learning_rate": 1.4679424607493315e-05,
+ "loss": 2.7540690898895264,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.026957856491208076,
+ "learning_rate": 1.4615563716290698e-05,
+ "loss": 2.774245500564575,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.028008051216602325,
+ "learning_rate": 1.455193517121951e-05,
+ "loss": 2.768254280090332,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.027576304972171783,
+ "learning_rate": 1.448853902354421e-05,
+ "loss": 2.7471680641174316,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.02723308838903904,
+ "learning_rate": 1.4425375324341827e-05,
+ "loss": 2.782475233078003,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.028015952557325363,
+ "learning_rate": 1.4362444124502309e-05,
+ "loss": 2.761064052581787,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.02686639130115509,
+ "learning_rate": 1.429974547472814e-05,
+ "loss": 2.780773162841797,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.027318131178617477,
+ "learning_rate": 1.4237279425534442e-05,
+ "loss": 2.790548801422119,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.027507461607456207,
+ "learning_rate": 1.4175046027249065e-05,
+ "loss": 2.7799487113952637,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.027209127321839333,
+ "learning_rate": 1.4113045330012324e-05,
+ "loss": 2.7879908084869385,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.027000298723578453,
+ "learning_rate": 1.4051277383776953e-05,
+ "loss": 2.785269260406494,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.026835331693291664,
+ "learning_rate": 1.398974223830839e-05,
+ "loss": 2.7634077072143555,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.027131548151373863,
+ "learning_rate": 1.3928439943184394e-05,
+ "loss": 2.780407667160034,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.027154795825481415,
+ "learning_rate": 1.3867370547795065e-05,
+ "loss": 2.7430217266082764,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.027125395834445953,
+ "learning_rate": 1.3806534101342891e-05,
+ "loss": 2.7663848400115967,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.02727346308529377,
+ "learning_rate": 1.3745930652842728e-05,
+ "loss": 2.7848246097564697,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.027327269315719604,
+ "learning_rate": 1.368556025112176e-05,
+ "loss": 2.7878546714782715,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.027070313692092896,
+ "learning_rate": 1.362542294481918e-05,
+ "loss": 2.768770933151245,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.026788033545017242,
+ "learning_rate": 1.356551878238656e-05,
+ "loss": 2.776149034500122,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.027060916647315025,
+ "learning_rate": 1.3505847812087724e-05,
+ "loss": 2.7886319160461426,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.027068400755524635,
+ "learning_rate": 1.3446410081998384e-05,
+ "loss": 2.750335216522217,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.02673197351396084,
+ "learning_rate": 1.3387205640006494e-05,
+ "loss": 2.796083450317383,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.026712248101830482,
+ "learning_rate": 1.332823453381196e-05,
+ "loss": 2.7682032585144043,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.02712504006922245,
+ "learning_rate": 1.3269496810926807e-05,
+ "loss": 2.790653705596924,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.027259107679128647,
+ "learning_rate": 1.3210992518674848e-05,
+ "loss": 2.7920985221862793,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.02679583430290222,
+ "learning_rate": 1.315272170419201e-05,
+ "loss": 2.7540664672851562,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.026917198672890663,
+ "learning_rate": 1.3094684414426073e-05,
+ "loss": 2.7794861793518066,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.02691952884197235,
+ "learning_rate": 1.3036880696136532e-05,
+ "loss": 2.7636754512786865,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.027165602892637253,
+ "learning_rate": 1.2979310595894882e-05,
+ "loss": 2.8022518157958984,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.0270492322742939,
+ "learning_rate": 1.2921974160084251e-05,
+ "loss": 2.763423204421997,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.0269774217158556,
+ "learning_rate": 1.2864871434899672e-05,
+ "loss": 2.7642290592193604,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.02735322155058384,
+ "learning_rate": 1.2808002466347654e-05,
+ "loss": 2.7912001609802246,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.026876460760831833,
+ "learning_rate": 1.2751367300246535e-05,
+ "loss": 2.8102259635925293,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.027073487639427185,
+ "learning_rate": 1.2694965982226301e-05,
+ "loss": 2.7805755138397217,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.026842350140213966,
+ "learning_rate": 1.2638798557728402e-05,
+ "loss": 2.80954909324646,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.027058839797973633,
+ "learning_rate": 1.2582865072006e-05,
+ "loss": 2.79652738571167,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.026913955807685852,
+ "learning_rate": 1.2527165570123586e-05,
+ "loss": 2.7777507305145264,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.026756104081869125,
+ "learning_rate": 1.2471700096957376e-05,
+ "loss": 2.7783961296081543,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.026751704514026642,
+ "learning_rate": 1.2416468697194767e-05,
+ "loss": 2.765263080596924,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.02720797434449196,
+ "learning_rate": 1.2361471415334795e-05,
+ "loss": 2.7831039428710938,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.027288692072033882,
+ "learning_rate": 1.2306708295687706e-05,
+ "loss": 2.780830144882202,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.027102869004011154,
+ "learning_rate": 1.2252179382375211e-05,
+ "loss": 2.7735419273376465,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.027081217616796494,
+ "learning_rate": 1.2197884719330281e-05,
+ "loss": 2.761964797973633,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.02710801735520363,
+ "learning_rate": 1.2143824350297104e-05,
+ "loss": 2.7670583724975586,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.027270127087831497,
+ "learning_rate": 1.2089998318831137e-05,
+ "loss": 2.786961078643799,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.02709820494055748,
+ "learning_rate": 1.2036406668299062e-05,
+ "loss": 2.796149969100952,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.027192605659365654,
+ "learning_rate": 1.198304944187868e-05,
+ "loss": 2.7829535007476807,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.026996217668056488,
+ "learning_rate": 1.1929926682558995e-05,
+ "loss": 2.801142930984497,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.027222832664847374,
+ "learning_rate": 1.1877038433139998e-05,
+ "loss": 2.773264169692993,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.027082091197371483,
+ "learning_rate": 1.1824384736232879e-05,
+ "loss": 2.8198862075805664,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.027131790295243263,
+ "learning_rate": 1.1771965634259653e-05,
+ "loss": 2.8299996852874756,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.027189496904611588,
+ "learning_rate": 1.1719781169453536e-05,
+ "loss": 2.7663733959198,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.02728065848350525,
+ "learning_rate": 1.1667831383858562e-05,
+ "loss": 2.8235511779785156,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.035003211349248886,
+ "learning_rate": 1.1616116319329808e-05,
+ "loss": 2.812906265258789,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 4.58276891708374,
+ "eval_runtime": 57.7425,
+ "eval_samples_per_second": 42.291,
+ "eval_steps_per_second": 1.334,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.026835810393095016,
+ "learning_rate": 1.1564636017533036e-05,
+ "loss": 2.7499475479125977,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.02696690894663334,
+ "learning_rate": 1.151339051994508e-05,
+ "loss": 2.7551589012145996,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.026555679738521576,
+ "learning_rate": 1.146237986785356e-05,
+ "loss": 2.753525733947754,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.02668551355600357,
+ "learning_rate": 1.141160410235674e-05,
+ "loss": 2.7452681064605713,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.027071513235569,
+ "learning_rate": 1.1361063264363776e-05,
+ "loss": 2.77901029586792,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.027035463601350784,
+ "learning_rate": 1.1310757394594514e-05,
+ "loss": 2.769394874572754,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.026944896206259727,
+ "learning_rate": 1.1260686533579481e-05,
+ "loss": 2.78621768951416,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.026308422908186913,
+ "learning_rate": 1.121085072165986e-05,
+ "loss": 2.746005058288574,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.02692459337413311,
+ "learning_rate": 1.1161249998987429e-05,
+ "loss": 2.7592382431030273,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.026701323688030243,
+ "learning_rate": 1.111188440552471e-05,
+ "loss": 2.77679181098938,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.02708185277879238,
+ "learning_rate": 1.106275398104455e-05,
+ "loss": 2.782973289489746,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.026750970631837845,
+ "learning_rate": 1.1013858765130468e-05,
+ "loss": 2.760465621948242,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.026963144540786743,
+ "learning_rate": 1.0965198797176513e-05,
+ "loss": 2.7820069789886475,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.02676742523908615,
+ "learning_rate": 1.0916774116387103e-05,
+ "loss": 2.7657270431518555,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.027025895193219185,
+ "learning_rate": 1.0868584761777125e-05,
+ "loss": 2.7613558769226074,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.027024701237678528,
+ "learning_rate": 1.0820630772171869e-05,
+ "loss": 2.7726471424102783,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.027291523292660713,
+ "learning_rate": 1.0772912186206987e-05,
+ "loss": 2.807295799255371,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.02704339101910591,
+ "learning_rate": 1.0725429042328542e-05,
+ "loss": 2.788114309310913,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.02788626216351986,
+ "learning_rate": 1.0678181378792794e-05,
+ "loss": 2.7785983085632324,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.02694671042263508,
+ "learning_rate": 1.0631169233666306e-05,
+ "loss": 2.75980281829834,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.02722201682627201,
+ "learning_rate": 1.0584392644826016e-05,
+ "loss": 2.7798352241516113,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.026429232209920883,
+ "learning_rate": 1.053785164995886e-05,
+ "loss": 2.746462821960449,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.027395036071538925,
+ "learning_rate": 1.0491546286562148e-05,
+ "loss": 2.7769758701324463,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.027433833107352257,
+ "learning_rate": 1.0445476591943259e-05,
+ "loss": 2.7528226375579834,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.026795532554388046,
+ "learning_rate": 1.0399642603219712e-05,
+ "loss": 2.7847824096679688,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.026764724403619766,
+ "learning_rate": 1.0354044357319129e-05,
+ "loss": 2.7527782917022705,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.026894114911556244,
+ "learning_rate": 1.0308681890979173e-05,
+ "loss": 2.7632064819335938,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.026841631159186363,
+ "learning_rate": 1.026355524074758e-05,
+ "loss": 2.755859851837158,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.02638275921344757,
+ "learning_rate": 1.0218664442982084e-05,
+ "loss": 2.7602035999298096,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.026950951665639877,
+ "learning_rate": 1.0174009533850368e-05,
+ "loss": 2.8016932010650635,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.02662278153002262,
+ "learning_rate": 1.0129590549330119e-05,
+ "loss": 2.7938461303710938,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.02673911489546299,
+ "learning_rate": 1.0085407525208865e-05,
+ "loss": 2.7740683555603027,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.02670123428106308,
+ "learning_rate": 1.0041460497084135e-05,
+ "loss": 2.771376132965088,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.026872053742408752,
+ "learning_rate": 9.997749500363172e-06,
+ "loss": 2.7445852756500244,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.026589123532176018,
+ "learning_rate": 9.954274570263191e-06,
+ "loss": 2.7571749687194824,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.02649911306798458,
+ "learning_rate": 9.911035741811153e-06,
+ "loss": 2.768310546875,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.0264277383685112,
+ "learning_rate": 9.868033049843787e-06,
+ "loss": 2.7388150691986084,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.02685769833624363,
+ "learning_rate": 9.825266529007601e-06,
+ "loss": 2.793008327484131,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.026673899963498116,
+ "learning_rate": 9.782736213758782e-06,
+ "loss": 2.8166663646698,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.026844056323170662,
+ "learning_rate": 9.740442138363317e-06,
+ "loss": 2.773449182510376,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.026580123230814934,
+ "learning_rate": 9.698384336896679e-06,
+ "loss": 2.7908055782318115,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.02664843201637268,
+ "learning_rate": 9.656562843244114e-06,
+ "loss": 2.7975826263427734,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.026540955528616905,
+ "learning_rate": 9.614977691100445e-06,
+ "loss": 2.7841620445251465,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.026925355195999146,
+ "learning_rate": 9.57362891397007e-06,
+ "loss": 2.7862730026245117,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.02668730728328228,
+ "learning_rate": 9.532516545167004e-06,
+ "loss": 2.7859320640563965,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.02650708518922329,
+ "learning_rate": 9.491640617814634e-06,
+ "loss": 2.793994426727295,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.0267381239682436,
+ "learning_rate": 9.451001164846092e-06,
+ "loss": 2.786421298980713,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.02687137760221958,
+ "learning_rate": 9.410598219003725e-06,
+ "loss": 2.752488851547241,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.02661444991827011,
+ "learning_rate": 9.370431812839556e-06,
+ "loss": 2.7778420448303223,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.026941899210214615,
+ "learning_rate": 9.330501978714857e-06,
+ "loss": 2.777524948120117,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.0268207136541605,
+ "learning_rate": 9.290808748800408e-06,
+ "loss": 2.7725157737731934,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.02680084854364395,
+ "learning_rate": 9.25135215507634e-06,
+ "loss": 2.7609915733337402,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.026696713641285896,
+ "learning_rate": 9.212132229332094e-06,
+ "loss": 2.7365245819091797,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.0265318863093853,
+ "learning_rate": 9.173149003166458e-06,
+ "loss": 2.7494451999664307,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.026489339768886566,
+ "learning_rate": 9.134402507987533e-06,
+ "loss": 2.7605128288269043,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.026866141706705093,
+ "learning_rate": 9.095892775012672e-06,
+ "loss": 2.7785377502441406,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.02668614685535431,
+ "learning_rate": 9.057619835268437e-06,
+ "loss": 2.7544281482696533,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.026771903038024902,
+ "learning_rate": 9.019583719590641e-06,
+ "loss": 2.7701380252838135,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.026549167931079865,
+ "learning_rate": 8.981784458624375e-06,
+ "loss": 2.7562901973724365,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.02666434273123741,
+ "learning_rate": 8.944222082823717e-06,
+ "loss": 2.786470890045166,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.026594869792461395,
+ "learning_rate": 8.906896622452055e-06,
+ "loss": 2.791085958480835,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.026912730187177658,
+ "learning_rate": 8.86980810758183e-06,
+ "loss": 2.7630703449249268,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.02666492760181427,
+ "learning_rate": 8.832956568094596e-06,
+ "loss": 2.771186351776123,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.02624955028295517,
+ "learning_rate": 8.796342033680928e-06,
+ "loss": 2.740208625793457,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.02678191289305687,
+ "learning_rate": 8.759964533840574e-06,
+ "loss": 2.7912135124206543,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.026293903589248657,
+ "learning_rate": 8.723824097882145e-06,
+ "loss": 2.7572717666625977,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.02646786905825138,
+ "learning_rate": 8.687920754923393e-06,
+ "loss": 2.7562999725341797,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.026905927807092667,
+ "learning_rate": 8.652254533890956e-06,
+ "loss": 2.7835705280303955,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.026924896985292435,
+ "learning_rate": 8.616825463520493e-06,
+ "loss": 2.7965805530548096,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.026362398639321327,
+ "learning_rate": 8.581633572356576e-06,
+ "loss": 2.737618923187256,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.026691535487771034,
+ "learning_rate": 8.546678888752629e-06,
+ "loss": 2.776054620742798,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.03468878194689751,
+ "learning_rate": 8.51196144087103e-06,
+ "loss": 2.7775015830993652,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 4.590993881225586,
+ "eval_runtime": 57.7996,
+ "eval_samples_per_second": 42.249,
+ "eval_steps_per_second": 1.332,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.026764800772070885,
+ "learning_rate": 8.477481256683005e-06,
+ "loss": 2.7140045166015625,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.026526780799031258,
+ "learning_rate": 8.44323836396864e-06,
+ "loss": 2.7631232738494873,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.026312852278351784,
+ "learning_rate": 8.409232790316763e-06,
+ "loss": 2.7715871334075928,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.026490503922104836,
+ "learning_rate": 8.375464563125094e-06,
+ "loss": 2.7789485454559326,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.02680407650768757,
+ "learning_rate": 8.341933709600102e-06,
+ "loss": 2.7781310081481934,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.026628121733665466,
+ "learning_rate": 8.308640256756916e-06,
+ "loss": 2.754638195037842,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.026735341176390648,
+ "learning_rate": 8.275584231419512e-06,
+ "loss": 2.7532246112823486,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.026857327669858932,
+ "learning_rate": 8.24276566022055e-06,
+ "loss": 2.753671646118164,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.026332298293709755,
+ "learning_rate": 8.21018456960132e-06,
+ "loss": 2.752696990966797,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.026672109961509705,
+ "learning_rate": 8.177840985811858e-06,
+ "loss": 2.8153340816497803,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.02677585557103157,
+ "learning_rate": 8.145734934910757e-06,
+ "loss": 2.7665958404541016,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.02687709592282772,
+ "learning_rate": 8.113866442765364e-06,
+ "loss": 2.7748546600341797,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.026170765981078148,
+ "learning_rate": 8.082235535051449e-06,
+ "loss": 2.759859561920166,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.026528827846050262,
+ "learning_rate": 8.050842237253533e-06,
+ "loss": 2.804813861846924,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.0264429971575737,
+ "learning_rate": 8.019686574664564e-06,
+ "loss": 2.765291213989258,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.026966489851474762,
+ "learning_rate": 7.988768572386171e-06,
+ "loss": 2.7834548950195312,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.02647687867283821,
+ "learning_rate": 7.958088255328415e-06,
+ "loss": 2.7787539958953857,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.026527805253863335,
+ "learning_rate": 7.927645648209837e-06,
+ "loss": 2.7454590797424316,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.026494361460208893,
+ "learning_rate": 7.897440775557506e-06,
+ "loss": 2.7637276649475098,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.0264204703271389,
+ "learning_rate": 7.867473661706946e-06,
+ "loss": 2.7916486263275146,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.02635893039405346,
+ "learning_rate": 7.837744330802136e-06,
+ "loss": 2.7680537700653076,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.026351723819971085,
+ "learning_rate": 7.808252806795446e-06,
+ "loss": 2.7503089904785156,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.026458928361535072,
+ "learning_rate": 7.778999113447673e-06,
+ "loss": 2.7462029457092285,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.026599271222949028,
+ "learning_rate": 7.749983274327994e-06,
+ "loss": 2.7531824111938477,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.02652306668460369,
+ "learning_rate": 7.721205312813921e-06,
+ "loss": 2.7552847862243652,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.026365090161561966,
+ "learning_rate": 7.69266525209132e-06,
+ "loss": 2.739084243774414,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.02636418491601944,
+ "learning_rate": 7.664363115154476e-06,
+ "loss": 2.72769832611084,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.026417085900902748,
+ "learning_rate": 7.636298924805872e-06,
+ "loss": 2.726992607116699,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.02646288275718689,
+ "learning_rate": 7.608472703656282e-06,
+ "loss": 2.7834439277648926,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.026408160105347633,
+ "learning_rate": 7.580884474124833e-06,
+ "loss": 2.773449659347534,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.026651300489902496,
+ "learning_rate": 7.55353425843888e-06,
+ "loss": 2.762749671936035,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.02692393772304058,
+ "learning_rate": 7.526422078633939e-06,
+ "loss": 2.7532958984375,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.026534877717494965,
+ "learning_rate": 7.4995479565538476e-06,
+ "loss": 2.7620131969451904,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.02664080075919628,
+ "learning_rate": 7.4729119138506044e-06,
+ "loss": 2.7795252799987793,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.02610137313604355,
+ "learning_rate": 7.446513971984366e-06,
+ "loss": 2.740626096725464,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.026373442262411118,
+ "learning_rate": 7.420354152223485e-06,
+ "loss": 2.809202194213867,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.026477070525288582,
+ "learning_rate": 7.394432475644402e-06,
+ "loss": 2.7980127334594727,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.026563815772533417,
+ "learning_rate": 7.368748963131854e-06,
+ "loss": 2.7525994777679443,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.026580089703202248,
+ "learning_rate": 7.343303635378537e-06,
+ "loss": 2.7786154747009277,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.026262899860739708,
+ "learning_rate": 7.31809651288524e-06,
+ "loss": 2.755073070526123,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.02683323062956333,
+ "learning_rate": 7.2931276159609475e-06,
+ "loss": 2.784115791320801,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.0266089029610157,
+ "learning_rate": 7.268396964722602e-06,
+ "loss": 2.757763385772705,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.0266721174120903,
+ "learning_rate": 7.243904579095308e-06,
+ "loss": 2.7321813106536865,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.026465903967618942,
+ "learning_rate": 7.219650478812065e-06,
+ "loss": 2.7546777725219727,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.02655668370425701,
+ "learning_rate": 7.195634683413996e-06,
+ "loss": 2.7383837699890137,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.02678738348186016,
+ "learning_rate": 7.171857212250225e-06,
+ "loss": 2.7965049743652344,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.02643296681344509,
+ "learning_rate": 7.148318084477766e-06,
+ "loss": 2.757577657699585,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.026728913187980652,
+ "learning_rate": 7.125017319061728e-06,
+ "loss": 2.7963476181030273,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.026841776445508003,
+ "learning_rate": 7.101954934775084e-06,
+ "loss": 2.761970043182373,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.026342229917645454,
+ "learning_rate": 7.079130950198802e-06,
+ "loss": 2.7694449424743652,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.026661433279514313,
+ "learning_rate": 7.05654538372171e-06,
+ "loss": 2.761528491973877,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.026377758011221886,
+ "learning_rate": 7.034198253540665e-06,
+ "loss": 2.7743632793426514,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.02674027532339096,
+ "learning_rate": 7.012089577660255e-06,
+ "loss": 2.809512138366699,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.026375900954008102,
+ "learning_rate": 6.990219373893128e-06,
+ "loss": 2.7861762046813965,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.0262554120272398,
+ "learning_rate": 6.968587659859665e-06,
+ "loss": 2.7638261318206787,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.026556264609098434,
+ "learning_rate": 6.9471944529881376e-06,
+ "loss": 2.744332790374756,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.027108121663331985,
+ "learning_rate": 6.926039770514686e-06,
+ "loss": 2.7774176597595215,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.026768099516630173,
+ "learning_rate": 6.905123629483277e-06,
+ "loss": 2.7910962104797363,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.02630891092121601,
+ "learning_rate": 6.884446046745608e-06,
+ "loss": 2.7563364505767822,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.026759762316942215,
+ "learning_rate": 6.8640070389613055e-06,
+ "loss": 2.7848401069641113,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.02670259028673172,
+ "learning_rate": 6.843806622597661e-06,
+ "loss": 2.7814881801605225,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.026632139459252357,
+ "learning_rate": 6.823844813929796e-06,
+ "loss": 2.759962797164917,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.026562802493572235,
+ "learning_rate": 6.804121629040597e-06,
+ "loss": 2.7739615440368652,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.026842210441827774,
+ "learning_rate": 6.7846370838206455e-06,
+ "loss": 2.740704298019409,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.02690100111067295,
+ "learning_rate": 6.765391193968287e-06,
+ "loss": 2.790417194366455,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.026587538421154022,
+ "learning_rate": 6.7463839749896335e-06,
+ "loss": 2.7733266353607178,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.026449866592884064,
+ "learning_rate": 6.727615442198394e-06,
+ "loss": 2.759732484817505,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.026533210650086403,
+ "learning_rate": 6.709085610716107e-06,
+ "loss": 2.749570608139038,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.02660883031785488,
+ "learning_rate": 6.690794495471878e-06,
+ "loss": 2.7633883953094482,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.02642008289694786,
+ "learning_rate": 6.672742111202543e-06,
+ "loss": 2.7738819122314453,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.026371262967586517,
+ "learning_rate": 6.654928472452569e-06,
+ "loss": 2.750501871109009,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.034971874207258224,
+ "learning_rate": 6.637353593574088e-06,
+ "loss": 2.773648977279663,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 4.597776889801025,
+ "eval_runtime": 57.7413,
+ "eval_samples_per_second": 42.292,
+ "eval_steps_per_second": 1.334,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.026444269344210625,
+ "learning_rate": 6.620017488726867e-06,
+ "loss": 2.785430431365967,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.026068098843097687,
+ "learning_rate": 6.602920171878265e-06,
+ "loss": 2.7389187812805176,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.02634287439286709,
+ "learning_rate": 6.5860616568033465e-06,
+ "loss": 2.7681965827941895,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.02599756047129631,
+ "learning_rate": 6.569441957084671e-06,
+ "loss": 2.7339017391204834,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.026426687836647034,
+ "learning_rate": 6.5530610861124e-06,
+ "loss": 2.7654523849487305,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.026443755254149437,
+ "learning_rate": 6.536919057084357e-06,
+ "loss": 2.7667272090911865,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.02631179802119732,
+ "learning_rate": 6.521015883005866e-06,
+ "loss": 2.751314878463745,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.026355480775237083,
+ "learning_rate": 6.5053515766898195e-06,
+ "loss": 2.773685932159424,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.026273800060153008,
+ "learning_rate": 6.489926150756674e-06,
+ "loss": 2.7847769260406494,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.026335887610912323,
+ "learning_rate": 6.4747396176344176e-06,
+ "loss": 2.766859531402588,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.026322951540350914,
+ "learning_rate": 6.459791989558542e-06,
+ "loss": 2.780479907989502,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.02634085901081562,
+ "learning_rate": 6.445083278572134e-06,
+ "loss": 2.7381904125213623,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.026212027296423912,
+ "learning_rate": 6.430613496525685e-06,
+ "loss": 2.7327303886413574,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.026233792304992676,
+ "learning_rate": 6.416382655077248e-06,
+ "loss": 2.7877025604248047,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.026520710438489914,
+ "learning_rate": 6.402390765692381e-06,
+ "loss": 2.7866268157958984,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.026276713237166405,
+ "learning_rate": 6.388637839644073e-06,
+ "loss": 2.76708722114563,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.026201969012618065,
+ "learning_rate": 6.375123888012812e-06,
+ "loss": 2.7421412467956543,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.026183459907770157,
+ "learning_rate": 6.361848921686523e-06,
+ "loss": 2.766507148742676,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.0260463934391737,
+ "learning_rate": 6.348812951360662e-06,
+ "loss": 2.761979103088379,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.02621057629585266,
+ "learning_rate": 6.33601598753802e-06,
+ "loss": 2.739781141281128,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.026393795385956764,
+ "learning_rate": 6.323458040528923e-06,
+ "loss": 2.7508270740509033,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.026347093284130096,
+ "learning_rate": 6.311139120451062e-06,
+ "loss": 2.743001937866211,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.02633446641266346,
+ "learning_rate": 6.299059237229567e-06,
+ "loss": 2.7588772773742676,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.026023060083389282,
+ "learning_rate": 6.287218400596961e-06,
+ "loss": 2.7275609970092773,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.026417123153805733,
+ "learning_rate": 6.27561662009321e-06,
+ "loss": 2.757868766784668,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.026473915204405785,
+ "learning_rate": 6.264253905065675e-06,
+ "loss": 2.7649319171905518,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.02639617957174778,
+ "learning_rate": 6.2531302646690866e-06,
+ "loss": 2.740844488143921,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.026274150237441063,
+ "learning_rate": 6.242245707865545e-06,
+ "loss": 2.771275520324707,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.026055404916405678,
+ "learning_rate": 6.23160024342455e-06,
+ "loss": 2.791905403137207,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.025836700573563576,
+ "learning_rate": 6.221193879922969e-06,
+ "loss": 2.7217962741851807,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.026319170370697975,
+ "learning_rate": 6.211026625745005e-06,
+ "loss": 2.7746083736419678,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.026230650022625923,
+ "learning_rate": 6.201098489082232e-06,
+ "loss": 2.770860433578491,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.026214411482214928,
+ "learning_rate": 6.191409477933554e-06,
+ "loss": 2.7712812423706055,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.02619052492082119,
+ "learning_rate": 6.18195960010528e-06,
+ "loss": 2.7662360668182373,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.026257462799549103,
+ "learning_rate": 6.172748863210954e-06,
+ "loss": 2.7349135875701904,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.026416735723614693,
+ "learning_rate": 6.163777274671552e-06,
+ "loss": 2.757458448410034,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.026388030499219894,
+ "learning_rate": 6.155044841715289e-06,
+ "loss": 2.744500160217285,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.026406629011034966,
+ "learning_rate": 6.146551571377715e-06,
+ "loss": 2.7591867446899414,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.026343751698732376,
+ "learning_rate": 6.1382974705017456e-06,
+ "loss": 2.759531021118164,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.02655695006251335,
+ "learning_rate": 6.130282545737535e-06,
+ "loss": 2.7539916038513184,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.026398885995149612,
+ "learning_rate": 6.1225068035425725e-06,
+ "loss": 2.743812084197998,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.026152826845645905,
+ "learning_rate": 6.114970250181616e-06,
+ "loss": 2.7331364154815674,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.026342187076807022,
+ "learning_rate": 6.107672891726724e-06,
+ "loss": 2.7509233951568604,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.02646729163825512,
+ "learning_rate": 6.1006147340572916e-06,
+ "loss": 2.797689437866211,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.02662649005651474,
+ "learning_rate": 6.093795782859917e-06,
+ "loss": 2.757904529571533,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.026698576286435127,
+ "learning_rate": 6.087216043628502e-06,
+ "loss": 2.7585411071777344,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.026497380807995796,
+ "learning_rate": 6.0808755216642146e-06,
+ "loss": 2.759650945663452,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.026584723964333534,
+ "learning_rate": 6.07477422207556e-06,
+ "loss": 2.766024351119995,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.026535525918006897,
+ "learning_rate": 6.06891214977818e-06,
+ "loss": 2.7463855743408203,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.026703383773565292,
+ "learning_rate": 6.063289309495051e-06,
+ "loss": 2.7799477577209473,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.026495758444070816,
+ "learning_rate": 6.057905705756421e-06,
+ "loss": 2.767401695251465,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.02615673840045929,
+ "learning_rate": 6.052761342899737e-06,
+ "loss": 2.7813711166381836,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.02638433314859867,
+ "learning_rate": 6.0478562250697185e-06,
+ "loss": 2.788975477218628,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.026382725685834885,
+ "learning_rate": 6.043190356218355e-06,
+ "loss": 2.7315516471862793,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.02635836973786354,
+ "learning_rate": 6.038763740104835e-06,
+ "loss": 2.756807327270508,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.02632402814924717,
+ "learning_rate": 6.034576380295586e-06,
+ "loss": 2.754215717315674,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.026373671367764473,
+ "learning_rate": 6.030628280164338e-06,
+ "loss": 2.7971506118774414,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.026456575840711594,
+ "learning_rate": 6.026919442891958e-06,
+ "loss": 2.7842516899108887,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.026110582053661346,
+ "learning_rate": 6.023449871466615e-06,
+ "loss": 2.7473106384277344,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.02640441618859768,
+ "learning_rate": 6.020219568683678e-06,
+ "loss": 2.766808271408081,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.0265651848167181,
+ "learning_rate": 6.0172285371456925e-06,
+ "loss": 2.804321527481079,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.026652604341506958,
+ "learning_rate": 6.014476779262568e-06,
+ "loss": 2.775193214416504,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.026517583057284355,
+ "learning_rate": 6.011964297251285e-06,
+ "loss": 2.784888744354248,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.02647361345589161,
+ "learning_rate": 6.0096910931361275e-06,
+ "loss": 2.7551116943359375,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.026397163048386574,
+ "learning_rate": 6.007657168748551e-06,
+ "loss": 2.747201442718506,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.02646862529218197,
+ "learning_rate": 6.005862525727277e-06,
+ "loss": 2.780600070953369,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.026226090267300606,
+ "learning_rate": 6.0043071655181986e-06,
+ "loss": 2.779751777648926,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.026499273255467415,
+ "learning_rate": 6.002991089374446e-06,
+ "loss": 2.7725064754486084,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.026265744119882584,
+ "learning_rate": 6.0019142983563875e-06,
+ "loss": 2.7508344650268555,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.026274647563695908,
+ "learning_rate": 6.001076793331555e-06,
+ "loss": 2.7637529373168945,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.026484506204724312,
+ "learning_rate": 6.000478574974689e-06,
+ "loss": 2.753441095352173,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.034853070974349976,
+ "learning_rate": 6.000119643767767e-06,
+ "loss": 2.804600477218628,
+ "step": 3600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.534806680666112e+18,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-3600/training_args.bin b/runs/l2r90-baseline/checkpoint-3600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-3600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-72/chat_template.jinja b/runs/l2r90-baseline/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-72/config.json b/runs/l2r90-baseline/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-72/generation_config.json b/runs/l2r90-baseline/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-72/model.safetensors b/runs/l2r90-baseline/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..fe49a450132d2f620e9b8e49b23a55ec5b7271dc
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:52490a039defecc0537c3f91d26e2ce9ddd6d0c081d6938213e70c85becb49de
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-72/optimizer.pt b/runs/l2r90-baseline/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..fa6f354bd02413d44b5a9e3adaa2228116294afc
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3eb623ba0676ffcdb34a9eb85048dcc4763cd55f4477c3b7b570e58e4079c863
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5631561581b3634362004f6c036145ee8e54d0f4
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b184f8fc9facd6e4a6513602952b43b9446784300c3ea2da72aefb029f93aac1
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..5c9399d17854873e90933eddfa9944d5e3d0b38f
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:937f07b3f415f537fbc809c70f15f64b8410cd6a07adbe7d246f5ca7359b3b49
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..8a986c37b53237a7bde101804a5602d204f737ba
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:07e3f7eacec877950b3dad99b106bf80da30203dda6159b06f954c15b7f5d367
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3ec45d94b0661de4e43586eec83b2caf40c79a8e
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2a5087a11a5f0a226f1607ec68740ea6baa2599559138de38da40f89dd666683
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-72/scheduler.pt b/runs/l2r90-baseline/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..43cf114590f1088fd438450226c42885d843da41
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:878ec53f5c2a524fc37af578b94a9025923faefea55f621be8e5643a79915d0f
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-72/tokenizer.json b/runs/l2r90-baseline/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-72/trainer_state.json b/runs/l2r90-baseline/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..37e1d12727703363c9e03304966f12832b9a9ffb
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+16,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-72/training_args.bin b/runs/l2r90-baseline/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856
diff --git a/runs/l2r90-baseline/checkpoint-720/chat_template.jinja b/runs/l2r90-baseline/checkpoint-720/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/l2r90-baseline/checkpoint-720/config.json b/runs/l2r90-baseline/checkpoint-720/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/l2r90-baseline/checkpoint-720/generation_config.json b/runs/l2r90-baseline/checkpoint-720/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/l2r90-baseline/checkpoint-720/model.safetensors b/runs/l2r90-baseline/checkpoint-720/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..2f79755d77882de972fec84f50946b1190fb0ef7
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:180e7d4278088669ee7f57d6a61aaac74684a8f4ff9a36b69aac4e203a960731
+size 583356232
diff --git a/runs/l2r90-baseline/checkpoint-720/optimizer.pt b/runs/l2r90-baseline/checkpoint-720/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ae1fd7129ce4b8110cee25739fcf91603e8e784b
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f3173a961075e4fff1e113d04b690bfda0179968cf320f4766a35ed5d96c476b
+size 1166825338
diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0977c0bf00c01a46d716db730a3c9bcc82905aaf
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a02d728a763af949adfec4bf1d121e1daf1ef655fd3b17a42c2399e0221768dd
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..4e5cd4f14d76f3fe2c0762093e6bb01d818def44
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d60087aa7d44f5751e6a709354a90b705404a21a9437680f20377cf12743341b
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1dbfc9354146f87ee95d6d46bdc97b42ae7b6f65
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/rng_state_2.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1c962bfcedafeb4fe574f63e648405b9b28a01333d49c391529b2c7a4fead463
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth b/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth
new file mode 100644
index 0000000000000000000000000000000000000000..0a96d4dec11fa27b14ffd309db2393d72f178feb
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/rng_state_3.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:478ec022631f64178ab44a8355b560b75c942f4cf348cc636a8cb37ec3bec6cc
+size 15024
diff --git a/runs/l2r90-baseline/checkpoint-720/scheduler.pt b/runs/l2r90-baseline/checkpoint-720/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..53133f8d1ca0990636972b84579aec1212febb2a
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea248f5cdc3b9f2d6793617acdaa11d93bab34c0a87ea94ea6aaabc148225da6
+size 1064
diff --git a/runs/l2r90-baseline/checkpoint-720/tokenizer.json b/runs/l2r90-baseline/checkpoint-720/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json b/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-720/trainer_state.json b/runs/l2r90-baseline/checkpoint-720/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..0d47dcf6191b4c779498a357f56ba06e8b8c1324
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/trainer_state.json
@@ -0,0 +1,5146 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.0,
+ "eval_steps": 500,
+ "global_step": 720,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.15997567772865295,
+ "learning_rate": 0.0,
+ "loss": 12.023977279663086,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.16017116606235504,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 12.022809982299805,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.1574592888355255,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 11.986297607421875,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.15066885948181152,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 11.91261100769043,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.14228598773479462,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 11.817420959472656,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.1410660743713379,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 11.70869255065918,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.13418471813201904,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 11.597742080688477,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.12223710119724274,
+ "learning_rate": 4.2e-05,
+ "loss": 11.500282287597656,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.11431518942117691,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 11.406240463256836,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.11069707572460175,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 11.324350357055664,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10802163928747177,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 11.263275146484375,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.10728747397661209,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 11.206270217895508,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10647646337747574,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 11.158624649047852,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10614501684904099,
+ "learning_rate": 7.8e-05,
+ "loss": 11.114002227783203,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10632237792015076,
+ "learning_rate": 8.4e-05,
+ "loss": 11.067373275756836,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.1061672493815422,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.015380859375,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.10657714307308197,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 10.958027839660645,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.10590068250894547,
+ "learning_rate": 0.000102,
+ "loss": 10.900931358337402,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.10591627657413483,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 10.8344144821167,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.10607340931892395,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 10.761369705200195,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.10447441786527634,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 10.6970796585083,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.10463859140872955,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 10.612152099609375,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.1049257144331932,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 10.5253324508667,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.10448617488145828,
+ "learning_rate": 0.000138,
+ "loss": 10.441542625427246,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.10413316637277603,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 10.353134155273438,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.10503070801496506,
+ "learning_rate": 0.00015,
+ "loss": 10.251943588256836,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.10336724668741226,
+ "learning_rate": 0.000156,
+ "loss": 10.17003059387207,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.10526089370250702,
+ "learning_rate": 0.000162,
+ "loss": 10.058032989501953,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.10343154519796371,
+ "learning_rate": 0.000168,
+ "loss": 9.968986511230469,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.10399148613214493,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 9.858746528625488,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.1026586964726448,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 9.766483306884766,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.10253847390413284,
+ "learning_rate": 0.000186,
+ "loss": 9.65650463104248,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.10166377574205399,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 9.547554016113281,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.10167674720287323,
+ "learning_rate": 0.000198,
+ "loss": 9.438922882080078,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.09929938614368439,
+ "learning_rate": 0.000204,
+ "loss": 9.350300788879395,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.09858057647943497,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 9.234173774719238,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.09838391095399857,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 9.121601104736328,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.0974399745464325,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 9.014242172241211,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.09716248512268066,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 8.907164573669434,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.09500741958618164,
+ "learning_rate": 0.000234,
+ "loss": 8.81848430633545,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.09266163408756256,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 8.729146003723145,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.0910920575261116,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 8.625696182250977,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.0894983559846878,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 8.51424503326416,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.08491404354572296,
+ "learning_rate": 0.000258,
+ "loss": 8.461417198181152,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.0824626088142395,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 8.371938705444336,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.0831412672996521,
+ "learning_rate": 0.00027,
+ "loss": 8.28670597076416,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.07819175720214844,
+ "learning_rate": 0.000276,
+ "loss": 8.186386108398438,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.07501859962940216,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 8.095057487487793,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.0688866451382637,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 8.032234191894531,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.06329295784235,
+ "learning_rate": 0.000294,
+ "loss": 7.964580059051514,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.06046910956501961,
+ "learning_rate": 0.0003,
+ "loss": 7.915485858917236,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.050732146948575974,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 7.86663818359375,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.04609096050262451,
+ "learning_rate": 0.000312,
+ "loss": 7.804510116577148,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.03886866196990013,
+ "learning_rate": 0.000318,
+ "loss": 7.783796310424805,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.034636761993169785,
+ "learning_rate": 0.000324,
+ "loss": 7.746722221374512,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.02844400331377983,
+ "learning_rate": 0.00033,
+ "loss": 7.712285041809082,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.026054631918668747,
+ "learning_rate": 0.000336,
+ "loss": 7.708642482757568,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.0225310567766428,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 7.6868672370910645,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.018994266167283058,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 7.6628804206848145,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.017860861495137215,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 7.652166366577148,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.019245322793722153,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 7.612408638000488,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.01606360450387001,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 7.627114295959473,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.01609027571976185,
+ "learning_rate": 0.000372,
+ "loss": 7.614614486694336,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.014387412928044796,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 7.593182563781738,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.016893157735466957,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 7.60258674621582,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.01590251922607422,
+ "learning_rate": 0.00039,
+ "loss": 7.590658187866211,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.012655598111450672,
+ "learning_rate": 0.000396,
+ "loss": 7.571450233459473,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.013261383399367332,
+ "learning_rate": 0.000402,
+ "loss": 7.5402021408081055,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.01822199858725071,
+ "learning_rate": 0.000408,
+ "loss": 7.541982650756836,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.0237593874335289,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 7.508282661437988,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.05531860142946243,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 7.518495082855225,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.12719468772411346,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 7.575924873352051,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.544726848602295,
+ "eval_runtime": 56.6032,
+ "eval_samples_per_second": 43.142,
+ "eval_steps_per_second": 1.36,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.06889434158802032,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 7.532279968261719,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 7.538493633270264,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.02140442468225956,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 7.478395462036133,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.03172273188829422,
+ "learning_rate": 0.00045,
+ "loss": 7.495583534240723,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.04667133465409279,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 7.470980644226074,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.05474933981895447,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 7.445516109466553,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.01863257959485054,
+ "learning_rate": 0.000468,
+ "loss": 7.463066101074219,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.037733182311058044,
+ "learning_rate": 0.000474,
+ "loss": 7.457036018371582,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.026289211586117744,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 7.440082550048828,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.027869857847690582,
+ "learning_rate": 0.000486,
+ "loss": 7.412266731262207,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.021711770445108414,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 7.384769439697266,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.03738119453191757,
+ "learning_rate": 0.000498,
+ "loss": 7.355679512023926,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.034208305180072784,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 7.343960285186768,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.020337563008069992,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 7.337154388427734,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.036635156720876694,
+ "learning_rate": 0.000516,
+ "loss": 7.334864139556885,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.0275711789727211,
+ "learning_rate": 0.000522,
+ "loss": 7.3356451988220215,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.022682661190629005,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 7.28330659866333,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.034273840487003326,
+ "learning_rate": 0.000534,
+ "loss": 7.287396430969238,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.02616060897707939,
+ "learning_rate": 0.00054,
+ "loss": 7.272294044494629,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.01899445243179798,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 7.239999771118164,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.020839497447013855,
+ "learning_rate": 0.000552,
+ "loss": 7.235346794128418,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.023729125037789345,
+ "learning_rate": 0.000558,
+ "loss": 7.1834821701049805,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.0318775475025177,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 7.198540687561035,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.015114367939531803,
+ "learning_rate": 0.00057,
+ "loss": 7.212254524230957,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.02599942311644554,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 7.164607048034668,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.029708007350564003,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 7.175436019897461,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.0463266558945179,
+ "learning_rate": 0.000588,
+ "loss": 7.173397541046143,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.06682766228914261,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 7.164839744567871,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.07400168478488922,
+ "learning_rate": 0.0006,
+ "loss": 7.173515319824219,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.04783684387803078,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 7.16644287109375,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.042767491191625595,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 7.145862579345703,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.028570281341671944,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 7.101897716522217,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.04605371132493019,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 7.111356735229492,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.02357972227036953,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 7.06209659576416,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.04726218059659004,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 7.099649429321289,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.02519933320581913,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 7.050154209136963,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.027398832142353058,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 7.028330326080322,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.027254166081547737,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 7.049069404602051,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.02054189145565033,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 7.0402045249938965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.021891873329877853,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 7.023792743682861,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.017824960872530937,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 7.002220153808594,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.022561997175216675,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 7.0106201171875,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.016903111711144447,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 6.970179557800293,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.02128630317747593,
+ "learning_rate": 0.000599973080557108,
+ "loss": 7.003076553344727,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.02123585157096386,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 6.9614105224609375,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.01661432348191738,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 6.955131530761719,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.01780594512820244,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 6.93811559677124,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.020178694278001785,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 6.902839660644531,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.033332180231809616,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 6.918459415435791,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.044249728322029114,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 6.946916103363037,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.04895849898457527,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 6.9353156089782715,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.024643104523420334,
+ "learning_rate": 0.000599936710690505,
+ "loss": 6.915379524230957,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.02990383468568325,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 6.898656845092773,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.033634357154369354,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 6.923366546630859,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.03218965232372284,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 6.923885345458984,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.02356622740626335,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 6.899965286254883,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.03445356339216232,
+ "learning_rate": 0.00059990620421714,
+ "loss": 6.883448600769043,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.028274044394493103,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 6.884991645812988,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.017005041241645813,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 6.8683953285217285,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.031443849205970764,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 6.832046031951904,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.029097655788064003,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 6.835471153259277,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.07707645744085312,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 6.857353210449219,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.09299153089523315,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 6.933018207550049,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.047855719923973083,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 6.8583083152771,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.07969525456428528,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 6.864906311035156,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.1365756392478943,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 6.903354644775391,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.1410154104232788,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 6.947896957397461,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.07029050588607788,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 6.910849571228027,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.06964805722236633,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 6.882598876953125,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.05349893495440483,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 6.836871147155762,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.05424615368247032,
+ "learning_rate": 0.000599788973374255,
+ "loss": 6.832496643066406,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.044589463621377945,
+ "learning_rate": 0.000599778806120077,
+ "loss": 6.8051371574401855,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.840139865875244,
+ "eval_runtime": 56.8976,
+ "eval_samples_per_second": 42.919,
+ "eval_steps_per_second": 1.353,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.04068000614643097,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 6.811692714691162,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.03961706906557083,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 6.788667678833008,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.03070618025958538,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 6.7994585037231445,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.036584168672561646,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 6.767686367034912,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.02327805571258068,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 6.784345626831055,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.028179455548524857,
+ "learning_rate": 0.000599712781599403,
+ "loss": 6.769927024841309,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.025651710107922554,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 6.7423295974731445,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.025366952642798424,
+ "learning_rate": 0.000599688860879549,
+ "loss": 6.736351013183594,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.023156974464654922,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 6.697917461395264,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.021937215700745583,
+ "learning_rate": 0.000599663984012462,
+ "loss": 6.687577247619629,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.020132191479206085,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 6.69163179397583,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.018260158598423004,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 6.673725128173828,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.017023146152496338,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 6.660531520843506,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.020211776718497276,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 6.6298699378967285,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.01748521998524666,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 6.626924514770508,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.023990269750356674,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 6.628607749938965,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.060859110206365585,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 6.625316619873047,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.12787196040153503,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 6.758873462677002,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.0820903554558754,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 6.788875579833984,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.08051387965679169,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 6.788204193115234,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.04531198740005493,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 6.734226703643799,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03978786990046501,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 6.685703277587891,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.0414281040430069,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 6.713569641113281,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.040948331356048584,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 6.695756435394287,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.04339255020022392,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 6.697275638580322,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.038956254720687866,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 6.6156110763549805,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.03434654325246811,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 6.639013290405273,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.0459795817732811,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 6.603760242462158,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.032164230942726135,
+ "learning_rate": 0.000599379982511273,
+ "loss": 6.6016459465026855,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.02468927390873432,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 6.589484214782715,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.02615884505212307,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 6.5663275718688965,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.02823791094124317,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 6.585080623626709,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.02933267503976822,
+ "learning_rate": 0.000599309205504528,
+ "loss": 6.549171447753906,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.020691825076937675,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 6.518960475921631,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.02271026186645031,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 6.519137382507324,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.028786325827240944,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 6.516097545623779,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.024719521403312683,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 6.495189666748047,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.02592875435948372,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 6.505522727966309,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.022884991019964218,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 6.508190155029297,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.01747789792716503,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 6.475917816162109,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.017053667455911636,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 6.474056243896484,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.0169149748980999,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 6.437000751495361,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.017995426431298256,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 6.416827201843262,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.023745683953166008,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 6.447346210479736,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.01796909235417843,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 6.426738739013672,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.014229115098714828,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 6.4247660636901855,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.020987126976251602,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 6.384764194488525,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.037671539932489395,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 6.393845558166504,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.06533686071634293,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 6.418580055236816,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.09679993242025375,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 6.455891132354736,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.08761037141084671,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 6.458295822143555,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.06243691220879555,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 6.4045023918151855,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.052615419030189514,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 6.393614292144775,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.04153689369559288,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 6.329081058502197,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.03536545857787132,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 6.382253170013428,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.03506210446357727,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 6.3689374923706055,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.028134305030107498,
+ "learning_rate": 0.000598804365316586,
+ "loss": 6.350457191467285,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.025324443355202675,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 6.3551506996154785,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.027925321832299232,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 6.342905044555664,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.02472207508981228,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 6.320896148681641,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.022138327360153198,
+ "learning_rate": 0.000598706872384039,
+ "loss": 6.326817512512207,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.02824472263455391,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 6.333435535430908,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.015822919085621834,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 6.287126541137695,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.02211311087012291,
+ "learning_rate": 0.000598631251036868,
+ "loss": 6.274896144866943,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.01805170439183712,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 6.27794075012207,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.02028656005859375,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 6.288621425628662,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.02157517708837986,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 6.297999382019043,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.028337396681308746,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 6.256705284118652,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.036196526139974594,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 6.259498596191406,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.04559934884309769,
+ "learning_rate": 0.000598473577921366,
+ "loss": 6.26982307434082,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.05263037979602814,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 6.276828765869141,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.046778593212366104,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 6.2884979248046875,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.268282413482666,
+ "eval_runtime": 57.2037,
+ "eval_samples_per_second": 42.69,
+ "eval_steps_per_second": 1.346,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.029683491215109825,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 6.2267165184021,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.03019959293305874,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 6.223328590393066,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.031647924333810806,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 6.228819370269775,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.025646064430475235,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 6.219881534576416,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.03026074543595314,
+ "learning_rate": 0.000598278794687186,
+ "loss": 6.172661304473877,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.027252312749624252,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 6.181373119354248,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.03526923805475235,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 6.194186210632324,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.03685254603624344,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 6.187342166900635,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.036119766533374786,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 6.200807571411133,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.036502398550510406,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 6.162562370300293,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.03517023101449013,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 6.149142265319824,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.023205652832984924,
+ "learning_rate": 0.00059807235435179,
+ "loss": 6.148383140563965,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.0221528522670269,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 6.143627643585205,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.024407457560300827,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 6.146297931671143,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.02728901617228985,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 6.114583969116211,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.039508312940597534,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 6.149223327636719,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.048040807247161865,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 6.122624397277832,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.056522514671087265,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 6.141239166259766,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.06174373999238014,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 6.152597427368164,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.07200063765048981,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 6.107697486877441,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.1010260134935379,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 6.191191673278809,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.06763981282711029,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 6.132532119750977,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.042127929627895355,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 6.165022850036621,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.03653500974178314,
+ "learning_rate": 0.000597691359743243,
+ "loss": 6.143064975738525,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.03675027936697006,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 6.108728408813477,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.03305647522211075,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 6.108780860900879,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.026183469220995903,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 6.088522911071777,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.02158971130847931,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 6.063396453857422,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.029356449842453003,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 6.10193395614624,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.020083140581846237,
+ "learning_rate": 0.000597488038559129,
+ "loss": 6.059609413146973,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.02572067640721798,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 6.08794641494751,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.02591501921415329,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 6.081871032714844,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.023668579757213593,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 6.077000617980957,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.02440894953906536,
+ "learning_rate": 0.000597347745466109,
+ "loss": 6.063971519470215,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.024253705516457558,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 6.08235502243042,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.024689998477697372,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 5.998595237731934,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.028521878644824028,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 5.979642868041992,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.030266916379332542,
+ "learning_rate": 0.000597203657966319,
+ "loss": 6.000593185424805,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.033342741429805756,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 6.004687309265137,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.03348678722977638,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 6.029177665710449,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.025689199566841125,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 5.98867130279541,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.027882622554898262,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 5.974981784820557,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.04662966728210449,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 5.979356288909912,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.07830756902694702,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 6.011981964111328,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.07885197550058365,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 6.019845008850098,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.0604281947016716,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 5.991020679473877,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.059781573712825775,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 5.999235153198242,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.04918011277914047,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 6.012720108032227,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.03718101978302002,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 5.986680507659912,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.03947242721915245,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 5.959175109863281,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.04288070648908615,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 5.976179122924805,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.03525253012776375,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 5.987687587738037,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.030190549790859222,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 5.965185642242432,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.027395015582442284,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 5.958373069763184,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.02688959613442421,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 5.959633827209473,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.02119498886168003,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 5.934293270111084,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.022015074267983437,
+ "learning_rate": 0.000596467483454833,
+ "loss": 5.914854049682617,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.021204788237810135,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 5.900177955627441,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.018889140337705612,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 5.873917579650879,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.018911773338913918,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 5.88484001159668,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.019570615142583847,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 5.886139392852783,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.02666759118437767,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 5.876997470855713,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.03528648987412453,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 5.8770318031311035,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.0549609512090683,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 5.86689567565918,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.07202128320932388,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 5.897315979003906,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.04940229654312134,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 5.914303779602051,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.05204314738512039,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 5.918832302093506,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.0608876571059227,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 5.898425579071045,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.0627935454249382,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 5.890061378479004,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.04895327612757683,
+ "learning_rate": 0.000595914592474791,
+ "loss": 5.885444164276123,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.04046280309557915,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 5.8441972732543945,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.045558422803878784,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 5.856559753417969,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 5.899378776550293,
+ "eval_runtime": 58.627,
+ "eval_samples_per_second": 41.653,
+ "eval_steps_per_second": 1.313,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.04504650831222534,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 5.841728210449219,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.036045704036951065,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 5.8451619148254395,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.029153067618608475,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 5.811412811279297,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.027814218774437904,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 5.848013877868652,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.025015871971845627,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 5.781689643859863,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.02065589465200901,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 5.783626079559326,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.023988133296370506,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 5.800251483917236,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.01968221738934517,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 5.7953410148620605,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.021222546696662903,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 5.765391826629639,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.018830647692084312,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 5.77407693862915,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.021495003253221512,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 5.756778717041016,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.02049877494573593,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 5.745931625366211,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.020010637119412422,
+ "learning_rate": 0.000595227087813793,
+ "loss": 5.755594253540039,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.030872922390699387,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 5.775322914123535,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.056580882519483566,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 5.771434783935547,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.07649239897727966,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 5.812885284423828,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.05227350816130638,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 5.755342960357666,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.0459996834397316,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 5.770654201507568,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.04197477549314499,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 5.7795915603637695,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.039317045360803604,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 5.7311177253723145,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.0334056094288826,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 5.737773895263672,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.02994806133210659,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 5.693323612213135,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.03594277426600456,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 5.7090654373168945,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.03271142393350601,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 5.691743850708008,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.030780376866459846,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 5.72518253326416,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.02948964200913906,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 5.725653648376465,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.024995718151330948,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 5.695929527282715,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.026229677721858025,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 5.717738628387451,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.02641691267490387,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 5.687970161437988,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.022786157205700874,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 5.651585578918457,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.023804830387234688,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 5.67464542388916,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.025330908596515656,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 5.6631011962890625,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.03142685443162918,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 5.651077747344971,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.04587113484740257,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 5.647086143493652,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.04738520458340645,
+ "learning_rate": 0.00059412296156686,
+ "loss": 5.701745986938477,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.028803404420614243,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 5.648317813873291,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.041121263056993484,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 5.647163391113281,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.05275499075651169,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 5.666481018066406,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.060963090509176254,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 5.645685195922852,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.05624202638864517,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 5.6632795333862305,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.07903645187616348,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 5.693739891052246,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.06570398062467575,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 5.675642013549805,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.05279144272208214,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 5.685312747955322,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.05221335217356682,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 5.665262222290039,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.044064246118068695,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 5.647260665893555,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.03285257890820503,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 5.631403923034668,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.03275465965270996,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 5.6139068603515625,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.0317506417632103,
+ "learning_rate": 0.000593417134927994,
+ "loss": 5.6341753005981445,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.0239992905408144,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 5.601463794708252,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.02451021410524845,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 5.581960201263428,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.02237604930996895,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 5.602834224700928,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.01903100311756134,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 5.604010105133057,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.018574781715869904,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 5.556118965148926,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.019580982625484467,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 5.594083786010742,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.02127210609614849,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 5.574706077575684,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.02134147845208645,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 5.57991361618042,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.02030216157436371,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 5.554903984069824,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.01813247986137867,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 5.550544261932373,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.021302592009305954,
+ "learning_rate": 0.000592789007481325,
+ "loss": 5.560070991516113,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.029217662289738655,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 5.517913818359375,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.03720416501164436,
+ "learning_rate": 0.000592671765466188,
+ "loss": 5.551721096038818,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.03990121930837631,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 5.5301032066345215,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.03713113069534302,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 5.5560712814331055,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.039228230714797974,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 5.531858444213867,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.030399464070796967,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 5.5186357498168945,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.0415034294128418,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 5.512445449829102,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.04447956383228302,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 5.491172790527344,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.03943555802106857,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 5.51191520690918,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.03963116183876991,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 5.513973236083984,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.04397537559270859,
+ "learning_rate": 0.000592132629452205,
+ "loss": 5.515983581542969,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.04135015606880188,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 5.465071201324463,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.03931731730699539,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 5.504716873168945,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 5.549652099609375,
+ "eval_runtime": 57.761,
+ "eval_samples_per_second": 42.278,
+ "eval_steps_per_second": 1.333,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.03958690166473389,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 5.459286689758301,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.03299890458583832,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 5.45464563369751,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.028040863573551178,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 5.444796562194824,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.028365720063447952,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 5.447852611541748,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.02822224237024784,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 5.453883171081543,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.036075398325920105,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 5.462671279907227,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.04683876410126686,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 5.443917274475098,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.05457588657736778,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 5.444374084472656,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.04250836744904518,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 5.435029983520508,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.029477844014763832,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 5.421862602233887,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.03673284500837326,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 5.430391311645508,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.030489861965179443,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 5.389314651489258,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.04037046805024147,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 5.425060272216797,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.0445217490196228,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 5.437565326690674,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.03885014355182648,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 5.410930633544922,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.0334964245557785,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 5.412092208862305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.03499286621809006,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 5.382216453552246,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.04171644523739815,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 5.387385368347168,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.057558078318834305,
+ "learning_rate": 0.000590801330015633,
+ "loss": 5.437925815582275,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.053846780210733414,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 5.41644811630249,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.027259036898612976,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 5.409185409545898,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.03394043818116188,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 5.377973556518555,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.03731574863195419,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 5.382505893707275,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.04705141484737396,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 5.413661003112793,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.044811323285102844,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 5.417835235595703,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.04000752419233322,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 5.406160354614258,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.038364164531230927,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 5.357950687408447,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.030500127002596855,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 5.344951629638672,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.030644729733467102,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 5.3599348068237305,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.037976790219545364,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 5.354084491729736,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.04106011986732483,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 5.363678455352783,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.04334832355380058,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 5.350959300994873,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.049939028918743134,
+ "learning_rate": 0.000589856958339686,
+ "loss": 5.3317155838012695,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.05300058051943779,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 5.353314399719238,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.05943184345960617,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 5.3683929443359375,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.05600789934396744,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 5.345516204833984,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.05392718315124512,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 5.291389465332031,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.037913475185632706,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 5.353423118591309,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.03362123295664787,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 5.379500389099121,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.032558735460042953,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 5.34043025970459,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.04074402526021004,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 5.303869247436523,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.04270484298467636,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 5.27297830581665,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.04604959115386009,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 5.313268184661865,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.04462246224284172,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 5.31167459487915,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.04966094344854355,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 5.316695690155029,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.04662073776125908,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 5.316898822784424,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.04380315542221069,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 5.285087585449219,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.03239579498767853,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 5.304283142089844,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.029690183699131012,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 5.305715084075928,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.03444967418909073,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 5.2707672119140625,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.03302565962076187,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 5.277251243591309,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.028477763757109642,
+ "learning_rate": 0.0005885028868537,
+ "loss": 5.295318603515625,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.022988801822066307,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 5.249057292938232,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.026115216314792633,
+ "learning_rate": 0.000588355515561353,
+ "loss": 5.270636081695557,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.02576332725584507,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 5.263096332550049,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.023978352546691895,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 5.276605606079102,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.020695513114333153,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 5.2141432762146,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01952219195663929,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 5.214289665222168,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.022633206099271774,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 5.236690998077393,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.023082347586750984,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 5.2092156410217285,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.019620463252067566,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 5.264451503753662,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.021504493430256844,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 5.23444938659668,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.025672364979982376,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 5.205289363861084,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.034436631947755814,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 5.2047648429870605,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.042389530688524246,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 5.192008018493652,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.03310716897249222,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 5.162734031677246,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.03029288910329342,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 5.212151527404785,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.04619310051202774,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 5.215909004211426,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.05618087574839592,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 5.210192680358887,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.03783707693219185,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 5.201597690582275,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.03721104562282562,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 5.188891410827637,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.033336155116558075,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 5.1598615646362305,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.2592363357543945,
+ "eval_runtime": 57.3343,
+ "eval_samples_per_second": 42.592,
+ "eval_steps_per_second": 1.343,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.03719741478562355,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 5.1646623611450195,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.04888090118765831,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 5.171182155609131,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.05073149502277374,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 5.183949947357178,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.04187731072306633,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 5.176133155822754,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.039445552974939346,
+ "learning_rate": 0.000586594769789753,
+ "loss": 5.148100852966309,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.035551708191633224,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 5.151512145996094,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.031175289303064346,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 5.102468967437744,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.03696789592504501,
+ "learning_rate": 0.000586356179990052,
+ "loss": 5.119902610778809,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.04836135357618332,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 5.15183162689209,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.05206914246082306,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 5.154870986938477,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.046368084847927094,
+ "learning_rate": 0.000586115535536849,
+ "loss": 5.130651473999023,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.07146432995796204,
+ "learning_rate": 0.000586034864430614,
+ "loss": 5.175804138183594,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.08809789270162582,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 5.229716777801514,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.08116360008716583,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 5.21910285949707,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.08282570540904999,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 5.2336506843566895,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.07403547316789627,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 5.251604080200195,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.07024004310369492,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 5.267114162445068,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.056330062448978424,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 5.201976299285889,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.051763955503702164,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 5.1733198165893555,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.039900992065668106,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 5.222941875457764,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.04187863692641258,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 5.165033340454102,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.042645812034606934,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 5.157554626464844,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.03856075555086136,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 5.184629440307617,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.029387330636382103,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 5.118739128112793,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.027584781870245934,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 5.100855350494385,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.023655924946069717,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 5.154520034790039,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.024163050577044487,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 5.105189323425293,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.020369920879602432,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 5.113671779632568,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.021030737087130547,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 5.0847649574279785,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.01729249767959118,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 5.053225517272949,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.018704712390899658,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 5.079106330871582,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.02049064077436924,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 5.044185161590576,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.020736856386065483,
+ "learning_rate": 0.000584288196583439,
+ "loss": 5.041932106018066,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.016543276607990265,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 5.032522201538086,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.018414106220006943,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 5.037397384643555,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.01611698977649212,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 5.038179397583008,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.014310561120510101,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 5.0146484375,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.015472214668989182,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 5.060323715209961,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.016373401507735252,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 5.032100677490234,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.018694665282964706,
+ "learning_rate": 0.000583683733328402,
+ "loss": 5.036993026733398,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.019977087154984474,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 4.991037368774414,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.023741086944937706,
+ "learning_rate": 0.000583508993216961,
+ "loss": 5.043832778930664,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.025061562657356262,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 5.016516208648682,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.020848384127020836,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 4.986289978027344,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.015086354687809944,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 4.9948625564575195,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.015096964314579964,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 4.996342658996582,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.016932200640439987,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 4.998758316040039,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.019936930388212204,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 4.959854602813721,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.027658682316541672,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 4.968629837036133,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.03299275413155556,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 4.994814872741699,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.033717650920152664,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 4.976295471191406,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.02969924360513687,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 4.969155788421631,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.02846827358007431,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 4.989721298217773,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.03670600801706314,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 4.988689422607422,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.045570675283670425,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 4.974407196044922,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.04564111679792404,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 4.966331481933594,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.04173210263252258,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 4.925197601318359,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.04228242486715317,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 4.965020656585693,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.05144782364368439,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 5.0022149085998535,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.049400657415390015,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 4.9635114669799805,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.043692708015441895,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 4.977723121643066,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.04891742020845413,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 4.996312141418457,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.035564981400966644,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 4.940932273864746,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.04679827764630318,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 4.983131408691406,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.05489890277385712,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 4.977827548980713,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.04640809819102287,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 4.954761505126953,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.04439452290534973,
+ "learning_rate": 0.00058124865204371,
+ "loss": 4.914865970611572,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.04587096720933914,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 4.949592590332031,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.044119734317064285,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 4.956068992614746,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.043133001774549484,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 4.993522644042969,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.04315278306603432,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 4.894291877746582,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.04819472134113312,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 4.958306789398193,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.023706436157227,
+ "eval_runtime": 57.4082,
+ "eval_samples_per_second": 42.537,
+ "eval_steps_per_second": 1.341,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.04403403028845787,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 4.884971618652344,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.041993748396635056,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 4.916353225708008,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.042302004992961884,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 4.9012556076049805,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.044500596821308136,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 4.915240287780762,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.03914736211299896,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 4.885915756225586,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.0361703559756279,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 4.829287528991699,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.03432302549481392,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 4.926834583282471,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.0317804180085659,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 4.8707170486450195,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.035261351615190506,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 4.897791862487793,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.03276067599654198,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 4.864504814147949,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.034757066518068314,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 4.890812873840332,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.033726081252098083,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 4.831684112548828,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.03074726276099682,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 4.8534111976623535,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.02215893566608429,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 4.8311238288879395,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.02132033370435238,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 4.842617034912109,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.020825816318392754,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 4.84535026550293,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.019334357231855392,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 4.839937686920166,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.017372334375977516,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 4.8115553855896,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.0190406683832407,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 4.832542419433594,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.017372574657201767,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 4.793632507324219,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.019171632826328278,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 4.776987552642822,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.02109278179705143,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 4.8216657638549805,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.026234494522213936,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 4.8109049797058105,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.0357641875743866,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 4.821109771728516,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.03627947345376015,
+ "learning_rate": 0.000578351407086301,
+ "loss": 4.810632228851318,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.032253146171569824,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 4.7930521965026855,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.03402210772037506,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 4.772500991821289,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.0321584977209568,
+ "learning_rate": 0.000578050665972623,
+ "loss": 4.781399726867676,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.030242707580327988,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 4.802542686462402,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.02361336164176464,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 4.771912574768066,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.028781302273273468,
+ "learning_rate": 0.000577747930429834,
+ "loss": 4.796924591064453,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.027042683213949203,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 4.745218276977539,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.02715984545648098,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 4.792944431304932,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.0314096100628376,
+ "learning_rate": 0.00057744320265311,
+ "loss": 4.7649993896484375,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.028749193996191025,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 4.7980546951293945,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.02655309997498989,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 4.746520519256592,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.027497397735714912,
+ "learning_rate": 0.000577136484852073,
+ "loss": 4.7490715980529785,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.0343393050134182,
+ "learning_rate": 0.000577033803741424,
+ "loss": 4.74436092376709,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.04444412514567375,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 4.755973815917969,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.0437556728720665,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 4.784368515014648,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.0369921550154686,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 4.753410339355469,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.05038141459226608,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 4.755491256713867,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.04511422663927078,
+ "learning_rate": 0.00057651708808768,
+ "loss": 4.760293006896973,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.048762883991003036,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 4.755958080291748,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.06749686598777771,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 4.802557468414307,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.06451250612735748,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 4.812925338745117,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.05436156690120697,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 4.807103157043457,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.05233754590153694,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 4.782449722290039,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.05519746616482735,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 4.808449745178223,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.0578567236661911,
+ "learning_rate": 0.000575784433093151,
+ "loss": 4.7676005363464355,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.0588264986872673,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 4.805377960205078,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.04225790873169899,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 4.748775005340576,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.0376368910074234,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 4.759791374206543,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.03830771893262863,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 4.729974746704102,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.03684520348906517,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 4.729854583740234,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.02983408235013485,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 4.78502082824707,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.02854904904961586,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 4.717472076416016,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.023819150403141975,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 4.717124938964844,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.02477158047258854,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 4.71781063079834,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.022002534940838814,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 4.7480669021606445,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.02133483812212944,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 4.693770408630371,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.022407138720154762,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 4.720734119415283,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.020850926637649536,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 4.745505332946777,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.020420582965016365,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 4.688610076904297,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.019014107063412666,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 4.673436164855957,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.017393039539456367,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 4.6941118240356445,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.017877329140901566,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 4.667474269866943,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.017749635502696037,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 4.6901092529296875,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.018746376037597656,
+ "learning_rate": 0.000573741597999996,
+ "loss": 4.680166721343994,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016260217875242233,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 4.6799821853637695,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015099814161658287,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 4.680666923522949,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.016539735719561577,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 4.639509201049805,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 4.768754482269287,
+ "eval_runtime": 58.3352,
+ "eval_samples_per_second": 41.862,
+ "eval_steps_per_second": 1.32,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.019733518362045288,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 4.635312080383301,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.022571036592125893,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 4.6041765213012695,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.022238552570343018,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 4.648976802825928,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.02115180902183056,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 4.623968124389648,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.021789928898215294,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 4.620028018951416,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.02593887411057949,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 4.591062545776367,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.0347842313349247,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 4.632149696350098,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.04119519516825676,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 4.6254191398620605,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.04895750433206558,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 4.621372222900391,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.05752161890268326,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 4.61965274810791,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.05115870013833046,
+ "learning_rate": 0.000572185929672906,
+ "loss": 4.63119649887085,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.05721089243888855,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 4.652647018432617,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.0646839588880539,
+ "learning_rate": 0.000571960215520345,
+ "loss": 4.642184257507324,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.0682854950428009,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 4.676589012145996,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.057402849197387695,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 4.6585283279418945,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.05885373800992966,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 4.671728134155273,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.05265705659985542,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 4.641249656677246,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.05417842045426369,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 4.661381721496582,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.06602369248867035,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 4.684137344360352,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.06004182621836662,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 4.625332832336426,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.04801511764526367,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 4.664791107177734,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.04380105435848236,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 4.664966583251953,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.03858727216720581,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 4.661867618560791,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.03240576758980751,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 4.5855326652526855,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.028561925515532494,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 4.637331008911133,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.026389246806502342,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 4.598666667938232,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.02484819106757641,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 4.601729393005371,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.02522977627813816,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 4.609833717346191,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.02589508146047592,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 4.574592590332031,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.023374728858470917,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 4.61362361907959,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.02465851791203022,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 4.639899253845215,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.02258715033531189,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 4.5656538009643555,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.018009789288043976,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 4.61449670791626,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.02121281623840332,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 4.553830623626709,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.018078701570630074,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 4.534579277038574,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.0169303547590971,
+ "learning_rate": 0.000569302387925215,
+ "loss": 4.533995628356934,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.016318142414093018,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 4.544438362121582,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.016664575785398483,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 4.5913286209106445,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.01641504280269146,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 4.520743370056152,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.017012735828757286,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 4.517453193664551,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.01672312617301941,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 4.5235915184021,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.016893865540623665,
+ "learning_rate": 0.000568590328960903,
+ "loss": 4.543200492858887,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.017530135810375214,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 4.54912805557251,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.018698599189519882,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 4.522650718688965,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.01984071172773838,
+ "learning_rate": 0.000568231409438027,
+ "loss": 4.546173095703125,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.023218289017677307,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 4.5305094718933105,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.028968270868062973,
+ "learning_rate": 0.000567991061134677,
+ "loss": 4.521440505981445,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.032370273023843765,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 4.54168701171875,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.033967748284339905,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 4.538717269897461,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.028757929801940918,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 4.517228603363037,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.029757825657725334,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 4.510310173034668,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.03581003099679947,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 4.506901264190674,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.03483985736966133,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 4.549284934997559,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.031016027554869652,
+ "learning_rate": 0.000567143121979179,
+ "loss": 4.530084609985352,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.029373526573181152,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 4.492218971252441,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.031661372631788254,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 4.502824783325195,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.027251387014985085,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 4.49193811416626,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.02263176068663597,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 4.512019157409668,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.02355228364467621,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 4.508917331695557,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.023297371342778206,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 4.497771263122559,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.019364114850759506,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 4.51470947265625,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.023899901658296585,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 4.48771858215332,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.02330983616411686,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 4.478496551513672,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.027252234518527985,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 4.497672080993652,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.030886363238096237,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 4.49789571762085,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.02515559457242489,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 4.490771293640137,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.020107125863432884,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 4.481879234313965,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.022076716646552086,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 4.494697570800781,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.025525815784931183,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 4.517453193664551,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.03005894646048546,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 4.453815460205078,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.030178502202033997,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 4.493922233581543,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.029249394312500954,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 4.453763008117676,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 4.607339382171631,
+ "eval_runtime": 57.9216,
+ "eval_samples_per_second": 42.16,
+ "eval_steps_per_second": 1.329,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.03362950682640076,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 4.430700302124023,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.04223684221506119,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 4.486915588378906,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.05192171037197113,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 4.467541694641113,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.042525868862867355,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 4.45252799987793,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.0406370609998703,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 4.414548873901367,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.044476233422756195,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 4.453492164611816,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.04687326401472092,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 4.477725028991699,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.04011611267924309,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 4.423356533050537,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.04971884936094284,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 4.4853515625,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.05863234028220177,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 4.466001033782959,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.057710036635398865,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 4.466723442077637,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.05155074968934059,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 4.4823198318481445,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.04535634070634842,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 4.471219062805176,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.045147854834795,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 4.476888179779053,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.03730619326233864,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 4.428932189941406,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.034599728882312775,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 4.47627592086792,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.032904356718063354,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 4.422671794891357,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.03356337174773216,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 4.456406593322754,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.035042595118284225,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 4.371820449829102,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.03269032761454582,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 4.402605056762695,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.028261031955480576,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 4.433136940002441,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.03301700949668884,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 4.388359546661377,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.031664177775382996,
+ "learning_rate": 0.000561968326197397,
+ "loss": 4.41859245300293,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.026281699538230896,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 4.406128883361816,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.023097911849617958,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 4.389717102050781,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.022349286824464798,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 4.408173561096191,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.021958226338028908,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 4.434830188751221,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.024551713839173317,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 4.386944770812988,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.028048967942595482,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 4.402279376983643,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02733907848596573,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 4.403069496154785,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.020858559757471085,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 4.406920909881592,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.021491149440407753,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 4.385799407958984,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.021495329216122627,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 4.4252495765686035,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.019708355888724327,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 4.388864994049072,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.01886274851858616,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 4.335962295532227,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.01955341547727585,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 4.373560905456543,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.018962012603878975,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 4.380022048950195,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.016663335263729095,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 4.421034336090088,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.01676873490214348,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 4.382775783538818,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.014379126951098442,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 4.41373872756958,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.015897583216428757,
+ "learning_rate": 0.000559585254906272,
+ "loss": 4.377284049987793,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.016513120383024216,
+ "learning_rate": 0.000559450892376025,
+ "loss": 4.41200065612793,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.016625918447971344,
+ "learning_rate": 0.000559316323227926,
+ "loss": 4.381187438964844,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.016028188169002533,
+ "learning_rate": 0.000559181547570395,
+ "loss": 4.353909015655518,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.01682914048433304,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 4.382326126098633,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.017078889533877373,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 4.334949493408203,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.018182938918471336,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 4.348130226135254,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.01940886303782463,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 4.339404106140137,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.020644692704081535,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 4.358404159545898,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02433430217206478,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 4.3660502433776855,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.027824537828564644,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 4.34050178527832,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.02973582223057747,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 4.348322868347168,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.030396128073334694,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 4.356224060058594,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.029760530218482018,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 4.357470989227295,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.02855013869702816,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 4.335353374481201,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.03376635164022446,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 4.331794738769531,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.034431129693984985,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 4.336880207061768,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.03524135425686836,
+ "learning_rate": 0.000557273065985207,
+ "loss": 4.31537389755249,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.03152058273553848,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 4.37662410736084,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.02878640592098236,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 4.292388916015625,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.029663654044270515,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 4.344346523284912,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.027895759791135788,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 4.365072727203369,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.026465769857168198,
+ "learning_rate": 0.000556581718698499,
+ "loss": 4.334959506988525,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.03191852569580078,
+ "learning_rate": 0.000556442835876097,
+ "loss": 4.361069679260254,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.03535444661974907,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 4.349765777587891,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.03451581299304962,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 4.343142509460449,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.029810845851898193,
+ "learning_rate": 0.000556024962691404,
+ "loss": 4.307485580444336,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.027523217722773552,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 4.366406440734863,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03043208085000515,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 4.356929779052734,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.03058922104537487,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 4.348942279815674,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.032052166759967804,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 4.3496809005737305,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.03170694038271904,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 4.315959930419922,
+ "step": 720
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+17,
+ "train_batch_size": 8,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/l2r90-baseline/checkpoint-720/training_args.bin b/runs/l2r90-baseline/checkpoint-720/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..1ee534003ef080d9ce99775e26f510e83b1ad6c9
--- /dev/null
+++ b/runs/l2r90-baseline/checkpoint-720/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:561eeaca2f19e6a9a71ca959ffe1ee900a1ef425903fba31f850e4394274da85
+size 4856