diff --git a/.gitattributes b/.gitattributes
index 5de9a3a1339ab7801e096f1723c451e1391619f6..e40e7b09629db7f7978c51ee801e664487698ca4 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -96,3 +96,14 @@ runs/baseline-lr6e-4/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs
runs/baseline-lr6e-4/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/baseline-lr6e-4/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/baseline-lr6e-4/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/mt5-ssl/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/mt5-ssl/checkpoint-1080/chat_template.jinja b/runs/mt5-ssl/checkpoint-1080/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-1080/config.json b/runs/mt5-ssl/checkpoint-1080/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-1080/generation_config.json b/runs/mt5-ssl/checkpoint-1080/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-1080/model.safetensors b/runs/mt5-ssl/checkpoint-1080/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4e3b9a5adc7cd4ab606c040add07572bcf416c27
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2b8b6da94ee5b243ae9a41f4e320d99dee014313dcd1742fab5fd7a7fca504ac
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-1080/optimizer.pt b/runs/mt5-ssl/checkpoint-1080/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e99b066455bcf2f70889ce92fbfd69a8273b250c
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f14e1df4b11fe90230a63988e77e58124b2daaad72f33036c24ce3350b7a9f8b
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-1080/rng_state_0.pth b/runs/mt5-ssl/checkpoint-1080/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ff81a4a04de793f1c690eab304502628a6f01a16
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3e64ff3db32365cabac6ffad743c22ab2586a3c7538842f851d54fcc9b00320b
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1080/rng_state_1.pth b/runs/mt5-ssl/checkpoint-1080/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2e0fc04a0aae1c0cb2027f4e52b0d2e25910d9a8
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:837b65fefb8e1f24a6981b0d4649ff164abbce25b5c9c0d885ed58d7e0986bae
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1080/scheduler.pt b/runs/mt5-ssl/checkpoint-1080/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3e275d97268eb212b051090ffcda70a20a7bd2eb
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bdeda49f9c209232e61e89d6b373289c81e91ac24ee327462ddf24dddafe25ed
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-1080/tokenizer.json b/runs/mt5-ssl/checkpoint-1080/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-1080/tokenizer_config.json b/runs/mt5-ssl/checkpoint-1080/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1080/trainer_state.json b/runs/mt5-ssl/checkpoint-1080/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..66048ad003e8b7cd7a214df6df450aeac99259bd
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/trainer_state.json
@@ -0,0 +1,7706 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.0,
+ "eval_steps": 500,
+ "global_step": 1080,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.6164840246411264e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1080/training_args.bin b/runs/mt5-ssl/checkpoint-1080/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1080/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-1440/chat_template.jinja b/runs/mt5-ssl/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-1440/config.json b/runs/mt5-ssl/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-1440/generation_config.json b/runs/mt5-ssl/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-1440/model.safetensors b/runs/mt5-ssl/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..632e435f41f17b5f9c1d564688f7652520319400
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26f9cceaee3d3b1c1b20b3f788a11060b31031b7272634ef3f783094bc0690cc
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-1440/optimizer.pt b/runs/mt5-ssl/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..fda57c470292f37fd493eb36aa7e0f680babf0e4
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e4e4595fceb3c71b9dba3eda9a45b8e8e32726dc882b2d159a0c7bc8c196da1e
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-1440/rng_state_0.pth b/runs/mt5-ssl/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3c396253e5ad58aa3254e1cb9364198815e3ba61
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad41999a3021e51cc92b222472cf4756a254132cabf58c4c2e1ee206c7366885
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1440/rng_state_1.pth b/runs/mt5-ssl/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..08b5f24cb1ec06dc192e471442f86bfe5e398030
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:29848a162fabe62f2066ebb397d9dc38254fb7b578fd1c8908975df46738482d
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1440/scheduler.pt b/runs/mt5-ssl/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e00c7bc69a3fbefc469c85cfede8a8a9403ecdc8
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5ca62ced3e9ed490d18d005ea4ac3ea68e1b051ea3ef0feded5dd8875ffb40a4
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-1440/tokenizer.json b/runs/mt5-ssl/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-1440/tokenizer_config.json b/runs/mt5-ssl/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1440/trainer_state.json b/runs/mt5-ssl/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2243ef3ff12d89e7a39ef232d5f6fc154c71637b
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.155312032854835e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1440/training_args.bin b/runs/mt5-ssl/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-1800/chat_template.jinja b/runs/mt5-ssl/checkpoint-1800/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-1800/config.json b/runs/mt5-ssl/checkpoint-1800/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-1800/generation_config.json b/runs/mt5-ssl/checkpoint-1800/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-1800/model.safetensors b/runs/mt5-ssl/checkpoint-1800/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ab8f6b8fa3fb142c3265cb025bf649ec1ed4fe66
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad0932cf7c26992bfb55b026276d85eda15108b8c7805038b036b0e1ecfbc776
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-1800/optimizer.pt b/runs/mt5-ssl/checkpoint-1800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..97bcb5839056de5127b9292a5418353517d2c610
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ddfef599a2dacceb7cdcdeb687e7bb18e2046533af07a027c2ce7575f1b0377c
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-1800/rng_state_0.pth b/runs/mt5-ssl/checkpoint-1800/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..304a7576d9bd060137258dba420a3e1ff00587c1
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:987694c44f0af3a8ae4e43e3677bef48a1cfa8e57eee3832ef223e9b8ad80516
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1800/rng_state_1.pth b/runs/mt5-ssl/checkpoint-1800/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..75486c47bfa9499bfe91975e3af3355c9526365c
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea5ecba78b648c2a45cab91d590e52b7295077372e6bf91c1591d3fb6a5535ca
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-1800/scheduler.pt b/runs/mt5-ssl/checkpoint-1800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..cc00d28bf1d3c152dd57b4e152eebb5a75268a48
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5c13118c758fe95682074ef807a1660af2ca020e5b32dd995760a77b78e786f0
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-1800/tokenizer.json b/runs/mt5-ssl/checkpoint-1800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-1800/tokenizer_config.json b/runs/mt5-ssl/checkpoint-1800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1800/trainer_state.json b/runs/mt5-ssl/checkpoint-1800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ef0690981a95d5dfce84f8a5e7b224d69020b62b
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/trainer_state.json
@@ -0,0 +1,12826 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.0,
+ "eval_steps": 500,
+ "global_step": 1800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.694140041068544e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-1800/training_args.bin b/runs/mt5-ssl/checkpoint-1800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-1800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-2160/chat_template.jinja b/runs/mt5-ssl/checkpoint-2160/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-2160/config.json b/runs/mt5-ssl/checkpoint-2160/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-2160/generation_config.json b/runs/mt5-ssl/checkpoint-2160/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-2160/model.safetensors b/runs/mt5-ssl/checkpoint-2160/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9b8f64138fe9bf9b7a236b135e00a19ba04b4430
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c539ba48db6a347f05c93d4ee2a2021968df0d3a1fef520bf855e2c7d814b090
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-2160/optimizer.pt b/runs/mt5-ssl/checkpoint-2160/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c4a01ef811ee230ab7210d8c0d5c9959f122a522
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:95f0b3a3942918dd5d0c694ac2add93529c20b85606ff202db12d096c8cb9997
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-2160/rng_state_0.pth b/runs/mt5-ssl/checkpoint-2160/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..86ca1aad69953afb19f9f8da4c01764c2f44d22b
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:253e65ee29b5e996242fd0eef4ab59e40fe385a114b29c442de150cf74084b60
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2160/rng_state_1.pth b/runs/mt5-ssl/checkpoint-2160/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3031af55ab1520df411513147e482ece04c4d4f6
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:04c9d382d23a06579a44352e657618aaf7b647b0f5de65751c81300c8f96f42a
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2160/scheduler.pt b/runs/mt5-ssl/checkpoint-2160/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..cfe439753906dadad51f11779ef9f425378a7857
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5cb2a89dd39217d5729496222633b6ae963778e9f10a9c38309eff07088a5195
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-2160/tokenizer.json b/runs/mt5-ssl/checkpoint-2160/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-2160/tokenizer_config.json b/runs/mt5-ssl/checkpoint-2160/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2160/trainer_state.json b/runs/mt5-ssl/checkpoint-2160/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..79f048129a08f04bc074ab0e2b94dee4caa92b26
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/trainer_state.json
@@ -0,0 +1,15386 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.0,
+ "eval_steps": 500,
+ "global_step": 2160,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6287577152252197,
+ "eval_runtime": 39.8508,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013937282229964,
+ "grad_norm": 0.056210409849882126,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 1.9758458137512207,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027874564459932,
+ "grad_norm": 0.06387963145971298,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 1.9743283987045288,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041811846689896,
+ "grad_norm": 0.059931688010692596,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 1.983224630355835,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05574912891986,
+ "grad_norm": 0.059698957949876785,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 1.9638850688934326,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069686411149824,
+ "grad_norm": 0.057556964457035065,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 1.9888463020324707,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083623693379792,
+ "grad_norm": 0.06250642985105515,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 1.9776880741119385,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097560975609756,
+ "grad_norm": 0.06218443810939789,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 1.9904823303222656,
+ "step": 1807
+ },
+ {
+ "epoch": 25.11149825783972,
+ "grad_norm": 0.06147411838173866,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 1.9881136417388916,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125435540069688,
+ "grad_norm": 0.06080890819430351,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 1.9833264350891113,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139372822299652,
+ "grad_norm": 0.057465583086013794,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 1.9764835834503174,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153310104529616,
+ "grad_norm": 0.061488356441259384,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 1.9807426929473877,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16724738675958,
+ "grad_norm": 0.060737013816833496,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 1.9796113967895508,
+ "step": 1812
+ },
+ {
+ "epoch": 25.181184668989548,
+ "grad_norm": 0.060508858412504196,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 1.9862518310546875,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195121951219512,
+ "grad_norm": 0.06062212213873863,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 1.9855456352233887,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209059233449477,
+ "grad_norm": 0.06143258139491081,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 1.9763805866241455,
+ "step": 1815
+ },
+ {
+ "epoch": 25.222996515679444,
+ "grad_norm": 0.06446489691734314,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 1.9674923419952393,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23693379790941,
+ "grad_norm": 0.06179432198405266,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 1.9792906045913696,
+ "step": 1817
+ },
+ {
+ "epoch": 25.250871080139373,
+ "grad_norm": 0.062219955027103424,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 1.983303189277649,
+ "step": 1818
+ },
+ {
+ "epoch": 25.264808362369337,
+ "grad_norm": 0.060631707310676575,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 1.9859293699264526,
+ "step": 1819
+ },
+ {
+ "epoch": 25.278745644599304,
+ "grad_norm": 0.06329842656850815,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 1.988373041152954,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29268292682927,
+ "grad_norm": 0.06290270388126373,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 1.9817759990692139,
+ "step": 1821
+ },
+ {
+ "epoch": 25.306620209059233,
+ "grad_norm": 0.062234487384557724,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 1.9615445137023926,
+ "step": 1822
+ },
+ {
+ "epoch": 25.320557491289197,
+ "grad_norm": 0.06063779070973396,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 1.9736756086349487,
+ "step": 1823
+ },
+ {
+ "epoch": 25.334494773519165,
+ "grad_norm": 0.06245938315987587,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 1.9932905435562134,
+ "step": 1824
+ },
+ {
+ "epoch": 25.34843205574913,
+ "grad_norm": 0.06380622833967209,
+ "learning_rate": 0.000309930621834911,
+ "loss": 1.9695440530776978,
+ "step": 1825
+ },
+ {
+ "epoch": 25.362369337979093,
+ "grad_norm": 0.0626329705119133,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 1.9820970296859741,
+ "step": 1826
+ },
+ {
+ "epoch": 25.37630662020906,
+ "grad_norm": 0.060844022780656815,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 1.9926050901412964,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390243902439025,
+ "grad_norm": 0.06239933893084526,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 2.001865863800049,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40418118466899,
+ "grad_norm": 0.06491564959287643,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 1.9897377490997314,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418118466898953,
+ "grad_norm": 0.05946580320596695,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 1.973170280456543,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43205574912892,
+ "grad_norm": 0.05930647253990173,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 1.9949123859405518,
+ "step": 1831
+ },
+ {
+ "epoch": 25.445993031358885,
+ "grad_norm": 0.06269685924053192,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 1.9912991523742676,
+ "step": 1832
+ },
+ {
+ "epoch": 25.45993031358885,
+ "grad_norm": 0.0637395903468132,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 1.9869414567947388,
+ "step": 1833
+ },
+ {
+ "epoch": 25.473867595818817,
+ "grad_norm": 0.0628778263926506,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 1.9748289585113525,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48780487804878,
+ "grad_norm": 0.06064455583691597,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 2.008216381072998,
+ "step": 1835
+ },
+ {
+ "epoch": 25.501742160278745,
+ "grad_norm": 0.0655415952205658,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 1.9852811098098755,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51567944250871,
+ "grad_norm": 0.06189798563718796,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 1.983716368675232,
+ "step": 1837
+ },
+ {
+ "epoch": 25.529616724738677,
+ "grad_norm": 0.061454080045223236,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 2.013745069503784,
+ "step": 1838
+ },
+ {
+ "epoch": 25.54355400696864,
+ "grad_norm": 0.06712181121110916,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 1.9856178760528564,
+ "step": 1839
+ },
+ {
+ "epoch": 25.557491289198605,
+ "grad_norm": 0.06364461034536362,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 2.0034379959106445,
+ "step": 1840
+ },
+ {
+ "epoch": 25.571428571428573,
+ "grad_norm": 0.06669855862855911,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 1.9868637323379517,
+ "step": 1841
+ },
+ {
+ "epoch": 25.585365853658537,
+ "grad_norm": 0.06578591465950012,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 1.9998185634613037,
+ "step": 1842
+ },
+ {
+ "epoch": 25.5993031358885,
+ "grad_norm": 0.06305381655693054,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 2.000594139099121,
+ "step": 1843
+ },
+ {
+ "epoch": 25.613240418118465,
+ "grad_norm": 0.061369556933641434,
+ "learning_rate": 0.000304866093757771,
+ "loss": 1.9811584949493408,
+ "step": 1844
+ },
+ {
+ "epoch": 25.627177700348433,
+ "grad_norm": 0.062390975654125214,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 2.0020859241485596,
+ "step": 1845
+ },
+ {
+ "epoch": 25.641114982578397,
+ "grad_norm": 0.06477145850658417,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 2.004000186920166,
+ "step": 1846
+ },
+ {
+ "epoch": 25.65505226480836,
+ "grad_norm": 0.06288585811853409,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 1.9919252395629883,
+ "step": 1847
+ },
+ {
+ "epoch": 25.66898954703833,
+ "grad_norm": 0.05959830805659294,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 1.9860963821411133,
+ "step": 1848
+ },
+ {
+ "epoch": 25.682926829268293,
+ "grad_norm": 0.0596182718873024,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 2.0014610290527344,
+ "step": 1849
+ },
+ {
+ "epoch": 25.696864111498257,
+ "grad_norm": 0.06665527820587158,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 2.0072989463806152,
+ "step": 1850
+ },
+ {
+ "epoch": 25.71080139372822,
+ "grad_norm": 0.0631684735417366,
+ "learning_rate": 0.000303,
+ "loss": 2.0076632499694824,
+ "step": 1851
+ },
+ {
+ "epoch": 25.72473867595819,
+ "grad_norm": 0.06214272975921631,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 2.0143537521362305,
+ "step": 1852
+ },
+ {
+ "epoch": 25.738675958188153,
+ "grad_norm": 0.06422685086727142,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 1.9885808229446411,
+ "step": 1853
+ },
+ {
+ "epoch": 25.752613240418118,
+ "grad_norm": 0.06583613157272339,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 2.0042953491210938,
+ "step": 1854
+ },
+ {
+ "epoch": 25.766550522648085,
+ "grad_norm": 0.06265029311180115,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 2.003157138824463,
+ "step": 1855
+ },
+ {
+ "epoch": 25.78048780487805,
+ "grad_norm": 0.0612902007997036,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 1.9832231998443604,
+ "step": 1856
+ },
+ {
+ "epoch": 25.794425087108014,
+ "grad_norm": 0.0625016912817955,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 2.0012407302856445,
+ "step": 1857
+ },
+ {
+ "epoch": 25.808362369337978,
+ "grad_norm": 0.062357913702726364,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 2.010338306427002,
+ "step": 1858
+ },
+ {
+ "epoch": 25.822299651567945,
+ "grad_norm": 0.06349784880876541,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 2.0105648040771484,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83623693379791,
+ "grad_norm": 0.061822179704904556,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 1.9987019300460815,
+ "step": 1860
+ },
+ {
+ "epoch": 25.850174216027874,
+ "grad_norm": 0.06443580985069275,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 1.994943618774414,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86411149825784,
+ "grad_norm": 0.061674103140830994,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 2.0116522312164307,
+ "step": 1862
+ },
+ {
+ "epoch": 25.878048780487806,
+ "grad_norm": 0.061944279819726944,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 2.012296438217163,
+ "step": 1863
+ },
+ {
+ "epoch": 25.89198606271777,
+ "grad_norm": 0.06270956248044968,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 2.0057339668273926,
+ "step": 1864
+ },
+ {
+ "epoch": 25.905923344947734,
+ "grad_norm": 0.06427973508834839,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 2.0053863525390625,
+ "step": 1865
+ },
+ {
+ "epoch": 25.9198606271777,
+ "grad_norm": 0.0637369304895401,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 2.0130667686462402,
+ "step": 1866
+ },
+ {
+ "epoch": 25.933797909407666,
+ "grad_norm": 0.06291835755109787,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 2.0094404220581055,
+ "step": 1867
+ },
+ {
+ "epoch": 25.94773519163763,
+ "grad_norm": 0.06374962627887726,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 2.0074503421783447,
+ "step": 1868
+ },
+ {
+ "epoch": 25.961672473867594,
+ "grad_norm": 0.06330475956201553,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 1.9999325275421143,
+ "step": 1869
+ },
+ {
+ "epoch": 25.975609756097562,
+ "grad_norm": 0.06077918782830238,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 1.994345784187317,
+ "step": 1870
+ },
+ {
+ "epoch": 25.989547038327526,
+ "grad_norm": 0.06524437665939331,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 2.0210394859313965,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.05064176768064499,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 1.4825201034545898,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6298588514328003,
+ "eval_runtime": 43.1768,
+ "eval_samples_per_second": 56.558,
+ "eval_steps_per_second": 0.463,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013937282229964,
+ "grad_norm": 0.05861504375934601,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 1.9728282690048218,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027874564459932,
+ "grad_norm": 0.06309985369443893,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 1.9638245105743408,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041811846689896,
+ "grad_norm": 0.06438111513853073,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 1.9697308540344238,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05574912891986,
+ "grad_norm": 0.060332708060741425,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 1.9574062824249268,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069686411149824,
+ "grad_norm": 0.061463311314582825,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 1.9469952583312988,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083623693379792,
+ "grad_norm": 0.06337480992078781,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 1.9695435762405396,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097560975609756,
+ "grad_norm": 0.06303273886442184,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 1.9764964580535889,
+ "step": 1879
+ },
+ {
+ "epoch": 26.11149825783972,
+ "grad_norm": 0.06061594933271408,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 1.9758808612823486,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125435540069688,
+ "grad_norm": 0.061181873083114624,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 1.9607892036437988,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139372822299652,
+ "grad_norm": 0.06557024270296097,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 1.9752333164215088,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153310104529616,
+ "grad_norm": 0.06261707097291946,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 1.9634147882461548,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16724738675958,
+ "grad_norm": 0.05993996560573578,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 1.957287311553955,
+ "step": 1884
+ },
+ {
+ "epoch": 26.181184668989548,
+ "grad_norm": 0.0612766332924366,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 1.9562265872955322,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195121951219512,
+ "grad_norm": 0.05735815688967705,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 1.9577300548553467,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209059233449477,
+ "grad_norm": 0.06258611381053925,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 1.9661660194396973,
+ "step": 1887
+ },
+ {
+ "epoch": 26.222996515679444,
+ "grad_norm": 0.06298062205314636,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 1.9508070945739746,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23693379790941,
+ "grad_norm": 0.06346822530031204,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 1.9677343368530273,
+ "step": 1889
+ },
+ {
+ "epoch": 26.250871080139373,
+ "grad_norm": 0.05959425866603851,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 1.9574894905090332,
+ "step": 1890
+ },
+ {
+ "epoch": 26.264808362369337,
+ "grad_norm": 0.0610729418694973,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 1.9515372514724731,
+ "step": 1891
+ },
+ {
+ "epoch": 26.278745644599304,
+ "grad_norm": 0.06362602859735489,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 1.966064453125,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29268292682927,
+ "grad_norm": 0.06124011054635048,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 1.9654541015625,
+ "step": 1893
+ },
+ {
+ "epoch": 26.306620209059233,
+ "grad_norm": 0.05773317068815231,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 1.9610934257507324,
+ "step": 1894
+ },
+ {
+ "epoch": 26.320557491289197,
+ "grad_norm": 0.05896785855293274,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 1.965133547782898,
+ "step": 1895
+ },
+ {
+ "epoch": 26.334494773519165,
+ "grad_norm": 0.059504859149456024,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 1.9617180824279785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.34843205574913,
+ "grad_norm": 0.05804497376084328,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 1.9761323928833008,
+ "step": 1897
+ },
+ {
+ "epoch": 26.362369337979093,
+ "grad_norm": 0.058211296796798706,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 1.9805936813354492,
+ "step": 1898
+ },
+ {
+ "epoch": 26.37630662020906,
+ "grad_norm": 0.05976752191781998,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 1.9514508247375488,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390243902439025,
+ "grad_norm": 0.05824121832847595,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 1.97285795211792,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40418118466899,
+ "grad_norm": 0.05793534219264984,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 1.9608209133148193,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418118466898953,
+ "grad_norm": 0.05697625130414963,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 1.9787089824676514,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43205574912892,
+ "grad_norm": 0.05703677609562874,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 1.9610021114349365,
+ "step": 1903
+ },
+ {
+ "epoch": 26.445993031358885,
+ "grad_norm": 0.057380311191082,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 1.9764081239700317,
+ "step": 1904
+ },
+ {
+ "epoch": 26.45993031358885,
+ "grad_norm": 0.05846763774752617,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 1.9641292095184326,
+ "step": 1905
+ },
+ {
+ "epoch": 26.473867595818817,
+ "grad_norm": 0.05929148569703102,
+ "learning_rate": 0.000288343693342466,
+ "loss": 1.9803684949874878,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48780487804878,
+ "grad_norm": 0.058236971497535706,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 1.981757402420044,
+ "step": 1907
+ },
+ {
+ "epoch": 26.501742160278745,
+ "grad_norm": 0.05845103785395622,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 1.9736125469207764,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51567944250871,
+ "grad_norm": 0.06104041263461113,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 1.9698188304901123,
+ "step": 1909
+ },
+ {
+ "epoch": 26.529616724738677,
+ "grad_norm": 0.0628688707947731,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 1.9786568880081177,
+ "step": 1910
+ },
+ {
+ "epoch": 26.54355400696864,
+ "grad_norm": 0.06115978956222534,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 1.9744799137115479,
+ "step": 1911
+ },
+ {
+ "epoch": 26.557491289198605,
+ "grad_norm": 0.06233362853527069,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 1.9801995754241943,
+ "step": 1912
+ },
+ {
+ "epoch": 26.571428571428573,
+ "grad_norm": 0.06182500347495079,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 1.9888432025909424,
+ "step": 1913
+ },
+ {
+ "epoch": 26.585365853658537,
+ "grad_norm": 0.05962767079472542,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 1.9648957252502441,
+ "step": 1914
+ },
+ {
+ "epoch": 26.5993031358885,
+ "grad_norm": 0.05994252488017082,
+ "learning_rate": 0.000285947841605349,
+ "loss": 1.9709041118621826,
+ "step": 1915
+ },
+ {
+ "epoch": 26.613240418118465,
+ "grad_norm": 0.06045769527554512,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 1.994588017463684,
+ "step": 1916
+ },
+ {
+ "epoch": 26.627177700348433,
+ "grad_norm": 0.059465229511260986,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 1.9743266105651855,
+ "step": 1917
+ },
+ {
+ "epoch": 26.641114982578397,
+ "grad_norm": 0.06020190939307213,
+ "learning_rate": 0.000285149463934261,
+ "loss": 1.9780995845794678,
+ "step": 1918
+ },
+ {
+ "epoch": 26.65505226480836,
+ "grad_norm": 0.05829813703894615,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 1.9831392765045166,
+ "step": 1919
+ },
+ {
+ "epoch": 26.66898954703833,
+ "grad_norm": 0.05960174649953842,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 1.9799871444702148,
+ "step": 1920
+ },
+ {
+ "epoch": 26.682926829268293,
+ "grad_norm": 0.06133116036653519,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 1.985158920288086,
+ "step": 1921
+ },
+ {
+ "epoch": 26.696864111498257,
+ "grad_norm": 0.05947015434503555,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 1.9783923625946045,
+ "step": 1922
+ },
+ {
+ "epoch": 26.71080139372822,
+ "grad_norm": 0.061687882989645004,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 1.9685261249542236,
+ "step": 1923
+ },
+ {
+ "epoch": 26.72473867595819,
+ "grad_norm": 0.0613567978143692,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 1.9945828914642334,
+ "step": 1924
+ },
+ {
+ "epoch": 26.738675958188153,
+ "grad_norm": 0.05902605876326561,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 1.966120719909668,
+ "step": 1925
+ },
+ {
+ "epoch": 26.752613240418118,
+ "grad_norm": 0.06018664687871933,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 1.991615653038025,
+ "step": 1926
+ },
+ {
+ "epoch": 26.766550522648085,
+ "grad_norm": 0.061171986162662506,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 1.9877536296844482,
+ "step": 1927
+ },
+ {
+ "epoch": 26.78048780487805,
+ "grad_norm": 0.06263386458158493,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 1.9799100160598755,
+ "step": 1928
+ },
+ {
+ "epoch": 26.794425087108014,
+ "grad_norm": 0.06326548755168915,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 1.9805757999420166,
+ "step": 1929
+ },
+ {
+ "epoch": 26.808362369337978,
+ "grad_norm": 0.06226228550076485,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 1.9867990016937256,
+ "step": 1930
+ },
+ {
+ "epoch": 26.822299651567945,
+ "grad_norm": 0.06003205478191376,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 1.9632174968719482,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83623693379791,
+ "grad_norm": 0.06075739488005638,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 1.9638378620147705,
+ "step": 1932
+ },
+ {
+ "epoch": 26.850174216027874,
+ "grad_norm": 0.061726443469524384,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 1.9865589141845703,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86411149825784,
+ "grad_norm": 0.0621049702167511,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 2.007071018218994,
+ "step": 1934
+ },
+ {
+ "epoch": 26.878048780487806,
+ "grad_norm": 0.061793744564056396,
+ "learning_rate": 0.000280627938758204,
+ "loss": 1.988924264907837,
+ "step": 1935
+ },
+ {
+ "epoch": 26.89198606271777,
+ "grad_norm": 0.062355563044548035,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 1.9830653667449951,
+ "step": 1936
+ },
+ {
+ "epoch": 26.905923344947734,
+ "grad_norm": 0.060881514102220535,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 2.003551959991455,
+ "step": 1937
+ },
+ {
+ "epoch": 26.9198606271777,
+ "grad_norm": 0.061703141778707504,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 1.9811534881591797,
+ "step": 1938
+ },
+ {
+ "epoch": 26.933797909407666,
+ "grad_norm": 0.0634910985827446,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 1.9840846061706543,
+ "step": 1939
+ },
+ {
+ "epoch": 26.94773519163763,
+ "grad_norm": 0.060807306319475174,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 1.9879541397094727,
+ "step": 1940
+ },
+ {
+ "epoch": 26.961672473867594,
+ "grad_norm": 0.06126011535525322,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 1.993988037109375,
+ "step": 1941
+ },
+ {
+ "epoch": 26.975609756097562,
+ "grad_norm": 0.06394185870885849,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 1.983365535736084,
+ "step": 1942
+ },
+ {
+ "epoch": 26.989547038327526,
+ "grad_norm": 0.06164516508579254,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 1.989990234375,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.04997039958834648,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 1.4921388626098633,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6310494542121887,
+ "eval_runtime": 40.875,
+ "eval_samples_per_second": 59.743,
+ "eval_steps_per_second": 0.489,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013937282229964,
+ "grad_norm": 0.05894223228096962,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 1.935828447341919,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027874564459932,
+ "grad_norm": 0.06731119751930237,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 1.945769190788269,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041811846689896,
+ "grad_norm": 0.06787799298763275,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 1.944061040878296,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05574912891986,
+ "grad_norm": 0.05840576812624931,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 1.94246506690979,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069686411149824,
+ "grad_norm": 0.058151841163635254,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 1.940029263496399,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083623693379792,
+ "grad_norm": 0.06329379975795746,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 1.9523427486419678,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097560975609756,
+ "grad_norm": 0.06530784070491791,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 1.9352362155914307,
+ "step": 1951
+ },
+ {
+ "epoch": 27.11149825783972,
+ "grad_norm": 0.06273649632930756,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 1.946252703666687,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125435540069688,
+ "grad_norm": 0.05996445193886757,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 1.9437150955200195,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139372822299652,
+ "grad_norm": 0.059087123721838,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 1.936135172843933,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153310104529616,
+ "grad_norm": 0.06257616728544235,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 1.9452544450759888,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16724738675958,
+ "grad_norm": 0.06591545045375824,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 1.9334893226623535,
+ "step": 1956
+ },
+ {
+ "epoch": 27.181184668989548,
+ "grad_norm": 0.0595872700214386,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 1.9514997005462646,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195121951219512,
+ "grad_norm": 0.05855429172515869,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 1.955148458480835,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209059233449477,
+ "grad_norm": 0.06383495032787323,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 1.9607688188552856,
+ "step": 1959
+ },
+ {
+ "epoch": 27.222996515679444,
+ "grad_norm": 0.06486682593822479,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 1.941009283065796,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23693379790941,
+ "grad_norm": 0.06293730437755585,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 1.9550879001617432,
+ "step": 1961
+ },
+ {
+ "epoch": 27.250871080139373,
+ "grad_norm": 0.06104227155447006,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 1.9483157396316528,
+ "step": 1962
+ },
+ {
+ "epoch": 27.264808362369337,
+ "grad_norm": 0.06920639425516129,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 1.948688268661499,
+ "step": 1963
+ },
+ {
+ "epoch": 27.278745644599304,
+ "grad_norm": 0.06186635419726372,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 1.9512240886688232,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29268292682927,
+ "grad_norm": 0.06258274614810944,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 1.9538028240203857,
+ "step": 1965
+ },
+ {
+ "epoch": 27.306620209059233,
+ "grad_norm": 0.06546445935964584,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 1.9497663974761963,
+ "step": 1966
+ },
+ {
+ "epoch": 27.320557491289197,
+ "grad_norm": 0.06289096921682358,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 1.9706182479858398,
+ "step": 1967
+ },
+ {
+ "epoch": 27.334494773519165,
+ "grad_norm": 0.06412214040756226,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 1.9423331022262573,
+ "step": 1968
+ },
+ {
+ "epoch": 27.34843205574913,
+ "grad_norm": 0.062773697078228,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 1.9471882581710815,
+ "step": 1969
+ },
+ {
+ "epoch": 27.362369337979093,
+ "grad_norm": 0.060290850698947906,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 1.9732742309570312,
+ "step": 1970
+ },
+ {
+ "epoch": 27.37630662020906,
+ "grad_norm": 0.062357380986213684,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 1.957550048828125,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390243902439025,
+ "grad_norm": 0.06070727854967117,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 1.951438069343567,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40418118466899,
+ "grad_norm": 0.06380726397037506,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 1.9512609243392944,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418118466898953,
+ "grad_norm": 0.06135795637965202,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 1.9663273096084595,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43205574912892,
+ "grad_norm": 0.0627567246556282,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 1.964203119277954,
+ "step": 1975
+ },
+ {
+ "epoch": 27.445993031358885,
+ "grad_norm": 0.06235141307115555,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 1.960247278213501,
+ "step": 1976
+ },
+ {
+ "epoch": 27.45993031358885,
+ "grad_norm": 0.06187676638364792,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 1.9667054414749146,
+ "step": 1977
+ },
+ {
+ "epoch": 27.473867595818817,
+ "grad_norm": 0.06090894341468811,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 1.9695777893066406,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48780487804878,
+ "grad_norm": 0.06363680213689804,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 1.9529361724853516,
+ "step": 1979
+ },
+ {
+ "epoch": 27.501742160278745,
+ "grad_norm": 0.06053898110985756,
+ "learning_rate": 0.00026868712586269,
+ "loss": 1.9707324504852295,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51567944250871,
+ "grad_norm": 0.061363156884908676,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 1.955542802810669,
+ "step": 1981
+ },
+ {
+ "epoch": 27.529616724738677,
+ "grad_norm": 0.06158224865794182,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 1.9489967823028564,
+ "step": 1982
+ },
+ {
+ "epoch": 27.54355400696864,
+ "grad_norm": 0.05886901170015335,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 1.9705257415771484,
+ "step": 1983
+ },
+ {
+ "epoch": 27.557491289198605,
+ "grad_norm": 0.05890485271811485,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 1.9430618286132812,
+ "step": 1984
+ },
+ {
+ "epoch": 27.571428571428573,
+ "grad_norm": 0.060433268547058105,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 1.9634191989898682,
+ "step": 1985
+ },
+ {
+ "epoch": 27.585365853658537,
+ "grad_norm": 0.0626354068517685,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 1.9568289518356323,
+ "step": 1986
+ },
+ {
+ "epoch": 27.5993031358885,
+ "grad_norm": 0.05973317474126816,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 1.9678292274475098,
+ "step": 1987
+ },
+ {
+ "epoch": 27.613240418118465,
+ "grad_norm": 0.062289632856845856,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 1.9413363933563232,
+ "step": 1988
+ },
+ {
+ "epoch": 27.627177700348433,
+ "grad_norm": 0.06162180379033089,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 1.9835751056671143,
+ "step": 1989
+ },
+ {
+ "epoch": 27.641114982578397,
+ "grad_norm": 0.06176188960671425,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 1.9541220664978027,
+ "step": 1990
+ },
+ {
+ "epoch": 27.65505226480836,
+ "grad_norm": 0.06479815393686295,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 1.9596972465515137,
+ "step": 1991
+ },
+ {
+ "epoch": 27.66898954703833,
+ "grad_norm": 0.06571854650974274,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 1.9628980159759521,
+ "step": 1992
+ },
+ {
+ "epoch": 27.682926829268293,
+ "grad_norm": 0.06167568638920784,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 1.9624769687652588,
+ "step": 1993
+ },
+ {
+ "epoch": 27.696864111498257,
+ "grad_norm": 0.06207694113254547,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 1.9487217664718628,
+ "step": 1994
+ },
+ {
+ "epoch": 27.71080139372822,
+ "grad_norm": 0.061430424451828,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 1.9567784070968628,
+ "step": 1995
+ },
+ {
+ "epoch": 27.72473867595819,
+ "grad_norm": 0.06589192897081375,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 1.9594354629516602,
+ "step": 1996
+ },
+ {
+ "epoch": 27.738675958188153,
+ "grad_norm": 0.06640033423900604,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 1.9637835025787354,
+ "step": 1997
+ },
+ {
+ "epoch": 27.752613240418118,
+ "grad_norm": 0.06181980296969414,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 1.9597129821777344,
+ "step": 1998
+ },
+ {
+ "epoch": 27.766550522648085,
+ "grad_norm": 0.06352445483207703,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 1.9612562656402588,
+ "step": 1999
+ },
+ {
+ "epoch": 27.78048780487805,
+ "grad_norm": 0.06398574262857437,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 1.9662930965423584,
+ "step": 2000
+ },
+ {
+ "epoch": 27.794425087108014,
+ "grad_norm": 0.06195084750652313,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 1.956035852432251,
+ "step": 2001
+ },
+ {
+ "epoch": 27.808362369337978,
+ "grad_norm": 0.05951351299881935,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 1.9712438583374023,
+ "step": 2002
+ },
+ {
+ "epoch": 27.822299651567945,
+ "grad_norm": 0.062185633927583694,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 1.9714539051055908,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83623693379791,
+ "grad_norm": 0.061178892850875854,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 1.9691438674926758,
+ "step": 2004
+ },
+ {
+ "epoch": 27.850174216027874,
+ "grad_norm": 0.06262431293725967,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 1.9506231546401978,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86411149825784,
+ "grad_norm": 0.06206081807613373,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 1.9662809371948242,
+ "step": 2006
+ },
+ {
+ "epoch": 27.878048780487806,
+ "grad_norm": 0.0626990869641304,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 1.972696304321289,
+ "step": 2007
+ },
+ {
+ "epoch": 27.89198606271777,
+ "grad_norm": 0.06345485150814056,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 1.9560344219207764,
+ "step": 2008
+ },
+ {
+ "epoch": 27.905923344947734,
+ "grad_norm": 0.06401040405035019,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 1.9732205867767334,
+ "step": 2009
+ },
+ {
+ "epoch": 27.9198606271777,
+ "grad_norm": 0.06437943130731583,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 1.974897027015686,
+ "step": 2010
+ },
+ {
+ "epoch": 27.933797909407666,
+ "grad_norm": 0.06296383589506149,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 1.9710102081298828,
+ "step": 2011
+ },
+ {
+ "epoch": 27.94773519163763,
+ "grad_norm": 0.06527025997638702,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 1.9707660675048828,
+ "step": 2012
+ },
+ {
+ "epoch": 27.961672473867594,
+ "grad_norm": 0.06185369938611984,
+ "learning_rate": 0.00025996500713765,
+ "loss": 1.9675114154815674,
+ "step": 2013
+ },
+ {
+ "epoch": 27.975609756097562,
+ "grad_norm": 0.06357269734144211,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 1.9812211990356445,
+ "step": 2014
+ },
+ {
+ "epoch": 27.989547038327526,
+ "grad_norm": 0.06581240892410278,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 1.97153639793396,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.049184102565050125,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 1.4647765159606934,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6328383088111877,
+ "eval_runtime": 40.5927,
+ "eval_samples_per_second": 60.159,
+ "eval_steps_per_second": 0.493,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013937282229964,
+ "grad_norm": 0.05999046564102173,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 1.9182333946228027,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027874564459932,
+ "grad_norm": 0.059804853051900864,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 1.924762487411499,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041811846689896,
+ "grad_norm": 0.05890650674700737,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 1.9341628551483154,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05574912891986,
+ "grad_norm": 0.057233694940805435,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 1.932476282119751,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069686411149824,
+ "grad_norm": 0.060101717710494995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 1.9403120279312134,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083623693379792,
+ "grad_norm": 0.06054912135004997,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 1.9274934530258179,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097560975609756,
+ "grad_norm": 0.05882031098008156,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 1.9331915378570557,
+ "step": 2023
+ },
+ {
+ "epoch": 28.11149825783972,
+ "grad_norm": 0.05846688151359558,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 1.9342529773712158,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125435540069688,
+ "grad_norm": 0.058974213898181915,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 1.9353152513504028,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139372822299652,
+ "grad_norm": 0.05882132798433304,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 1.9500072002410889,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153310104529616,
+ "grad_norm": 0.05919627472758293,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 1.9313960075378418,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16724738675958,
+ "grad_norm": 0.06413958966732025,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 1.9516873359680176,
+ "step": 2028
+ },
+ {
+ "epoch": 28.181184668989548,
+ "grad_norm": 0.05799189209938049,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 1.9290452003479004,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195121951219512,
+ "grad_norm": 0.05999152734875679,
+ "learning_rate": 0.000255486047794226,
+ "loss": 1.9396628141403198,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209059233449477,
+ "grad_norm": 0.060240138322114944,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 1.937988519668579,
+ "step": 2031
+ },
+ {
+ "epoch": 28.222996515679444,
+ "grad_norm": 0.05933740735054016,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 1.9343246221542358,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23693379790941,
+ "grad_norm": 0.05960093438625336,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 1.9326903820037842,
+ "step": 2033
+ },
+ {
+ "epoch": 28.250871080139373,
+ "grad_norm": 0.0635334923863411,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 1.9381543397903442,
+ "step": 2034
+ },
+ {
+ "epoch": 28.264808362369337,
+ "grad_norm": 0.06162877380847931,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 1.9265705347061157,
+ "step": 2035
+ },
+ {
+ "epoch": 28.278745644599304,
+ "grad_norm": 0.059690047055482864,
+ "learning_rate": 0.000253907826333243,
+ "loss": 1.9278099536895752,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29268292682927,
+ "grad_norm": 0.0612013153731823,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 1.939026117324829,
+ "step": 2037
+ },
+ {
+ "epoch": 28.306620209059233,
+ "grad_norm": 0.06080419942736626,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 1.9306461811065674,
+ "step": 2038
+ },
+ {
+ "epoch": 28.320557491289197,
+ "grad_norm": 0.06308622658252716,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 1.9350788593292236,
+ "step": 2039
+ },
+ {
+ "epoch": 28.334494773519165,
+ "grad_norm": 0.06124545633792877,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 1.9264733791351318,
+ "step": 2040
+ },
+ {
+ "epoch": 28.34843205574913,
+ "grad_norm": 0.059037938714027405,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 1.917725920677185,
+ "step": 2041
+ },
+ {
+ "epoch": 28.362369337979093,
+ "grad_norm": 0.06089222431182861,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 1.949327826499939,
+ "step": 2042
+ },
+ {
+ "epoch": 28.37630662020906,
+ "grad_norm": 0.05957084521651268,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 1.938978910446167,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390243902439025,
+ "grad_norm": 0.06060048192739487,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 1.9226362705230713,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40418118466899,
+ "grad_norm": 0.058304425328969955,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 1.9330644607543945,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418118466898953,
+ "grad_norm": 0.05970550701022148,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 1.931627631187439,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43205574912892,
+ "grad_norm": 0.05994785949587822,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 1.9387202262878418,
+ "step": 2047
+ },
+ {
+ "epoch": 28.445993031358885,
+ "grad_norm": 0.06001518666744232,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 1.951300859451294,
+ "step": 2048
+ },
+ {
+ "epoch": 28.45993031358885,
+ "grad_norm": 0.0636107549071312,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 1.9358865022659302,
+ "step": 2049
+ },
+ {
+ "epoch": 28.473867595818817,
+ "grad_norm": 0.059402115643024445,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 1.9298653602600098,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48780487804878,
+ "grad_norm": 0.06260179728269577,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 1.954633116722107,
+ "step": 2051
+ },
+ {
+ "epoch": 28.501742160278745,
+ "grad_norm": 0.06439191848039627,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 1.9338306188583374,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51567944250871,
+ "grad_norm": 0.06265310943126678,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 1.9500043392181396,
+ "step": 2053
+ },
+ {
+ "epoch": 28.529616724738677,
+ "grad_norm": 0.06323659420013428,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 1.9445586204528809,
+ "step": 2054
+ },
+ {
+ "epoch": 28.54355400696864,
+ "grad_norm": 0.06512445956468582,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 1.958587408065796,
+ "step": 2055
+ },
+ {
+ "epoch": 28.557491289198605,
+ "grad_norm": 0.06386571377515793,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 1.949887990951538,
+ "step": 2056
+ },
+ {
+ "epoch": 28.571428571428573,
+ "grad_norm": 0.06317705661058426,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 1.9512674808502197,
+ "step": 2057
+ },
+ {
+ "epoch": 28.585365853658537,
+ "grad_norm": 0.06504984200000763,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 1.9242960214614868,
+ "step": 2058
+ },
+ {
+ "epoch": 28.5993031358885,
+ "grad_norm": 0.06278593838214874,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 1.9203131198883057,
+ "step": 2059
+ },
+ {
+ "epoch": 28.613240418118465,
+ "grad_norm": 0.06171257048845291,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 1.954156517982483,
+ "step": 2060
+ },
+ {
+ "epoch": 28.627177700348433,
+ "grad_norm": 0.06303241103887558,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 1.9460458755493164,
+ "step": 2061
+ },
+ {
+ "epoch": 28.641114982578397,
+ "grad_norm": 0.06252635270357132,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 1.9477293491363525,
+ "step": 2062
+ },
+ {
+ "epoch": 28.65505226480836,
+ "grad_norm": 0.06226412579417229,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 1.9259647130966187,
+ "step": 2063
+ },
+ {
+ "epoch": 28.66898954703833,
+ "grad_norm": 0.06111626327037811,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 1.9573581218719482,
+ "step": 2064
+ },
+ {
+ "epoch": 28.682926829268293,
+ "grad_norm": 0.06460461020469666,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 1.9486393928527832,
+ "step": 2065
+ },
+ {
+ "epoch": 28.696864111498257,
+ "grad_norm": 0.062027834355831146,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 1.9404425621032715,
+ "step": 2066
+ },
+ {
+ "epoch": 28.71080139372822,
+ "grad_norm": 0.061942074447870255,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 1.9444060325622559,
+ "step": 2067
+ },
+ {
+ "epoch": 28.72473867595819,
+ "grad_norm": 0.06159934028983116,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 1.9635517597198486,
+ "step": 2068
+ },
+ {
+ "epoch": 28.738675958188153,
+ "grad_norm": 0.0621839202940464,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 1.9442765712738037,
+ "step": 2069
+ },
+ {
+ "epoch": 28.752613240418118,
+ "grad_norm": 0.06432677060365677,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 1.9604840278625488,
+ "step": 2070
+ },
+ {
+ "epoch": 28.766550522648085,
+ "grad_norm": 0.06037316843867302,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 1.948632836341858,
+ "step": 2071
+ },
+ {
+ "epoch": 28.78048780487805,
+ "grad_norm": 0.05999939888715744,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 1.9451751708984375,
+ "step": 2072
+ },
+ {
+ "epoch": 28.794425087108014,
+ "grad_norm": 0.06441698968410492,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 1.9356517791748047,
+ "step": 2073
+ },
+ {
+ "epoch": 28.808362369337978,
+ "grad_norm": 0.06059470772743225,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 1.9258331060409546,
+ "step": 2074
+ },
+ {
+ "epoch": 28.822299651567945,
+ "grad_norm": 0.062317874282598495,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 1.9342000484466553,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83623693379791,
+ "grad_norm": 0.06318258494138718,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 1.9395548105239868,
+ "step": 2076
+ },
+ {
+ "epoch": 28.850174216027874,
+ "grad_norm": 0.06204817444086075,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 1.9572374820709229,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86411149825784,
+ "grad_norm": 0.06250188499689102,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 1.9465739727020264,
+ "step": 2078
+ },
+ {
+ "epoch": 28.878048780487806,
+ "grad_norm": 0.06293781846761703,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 1.9431169033050537,
+ "step": 2079
+ },
+ {
+ "epoch": 28.89198606271777,
+ "grad_norm": 0.059349458664655685,
+ "learning_rate": 0.000242380656502223,
+ "loss": 1.9673197269439697,
+ "step": 2080
+ },
+ {
+ "epoch": 28.905923344947734,
+ "grad_norm": 0.06214660778641701,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 1.9646787643432617,
+ "step": 2081
+ },
+ {
+ "epoch": 28.9198606271777,
+ "grad_norm": 0.062279123812913895,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 1.9572803974151611,
+ "step": 2082
+ },
+ {
+ "epoch": 28.933797909407666,
+ "grad_norm": 0.06502588093280792,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 1.9510462284088135,
+ "step": 2083
+ },
+ {
+ "epoch": 28.94773519163763,
+ "grad_norm": 0.06153101101517677,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 1.9507923126220703,
+ "step": 2084
+ },
+ {
+ "epoch": 28.961672473867594,
+ "grad_norm": 0.06469229608774185,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 1.937751293182373,
+ "step": 2085
+ },
+ {
+ "epoch": 28.975609756097562,
+ "grad_norm": 0.0672568678855896,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 1.944124698638916,
+ "step": 2086
+ },
+ {
+ "epoch": 28.989547038327526,
+ "grad_norm": 0.06451335549354553,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 1.9519702196121216,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.05141456052660942,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 1.4445865154266357,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6343450546264648,
+ "eval_runtime": 40.5081,
+ "eval_samples_per_second": 60.284,
+ "eval_steps_per_second": 0.494,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013937282229964,
+ "grad_norm": 0.061562713235616684,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 1.921946406364441,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027874564459932,
+ "grad_norm": 0.06396839767694473,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 1.9222416877746582,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041811846689896,
+ "grad_norm": 0.05828692391514778,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 1.9173905849456787,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05574912891986,
+ "grad_norm": 0.06567556411027908,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 1.9148579835891724,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069686411149824,
+ "grad_norm": 0.05924173817038536,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 1.916333794593811,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083623693379792,
+ "grad_norm": 0.06129243224859238,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 1.9031038284301758,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097560975609756,
+ "grad_norm": 0.06068028509616852,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 1.9165983200073242,
+ "step": 2095
+ },
+ {
+ "epoch": 29.11149825783972,
+ "grad_norm": 0.06304074823856354,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 1.90432870388031,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125435540069688,
+ "grad_norm": 0.060179345309734344,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 1.9187151193618774,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139372822299652,
+ "grad_norm": 0.05633116513490677,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 1.9125087261199951,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153310104529616,
+ "grad_norm": 0.0598532073199749,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 1.906031847000122,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16724738675958,
+ "grad_norm": 0.05987119674682617,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 1.9177607297897339,
+ "step": 2100
+ },
+ {
+ "epoch": 29.181184668989548,
+ "grad_norm": 0.060883522033691406,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 1.9008433818817139,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195121951219512,
+ "grad_norm": 0.05977628752589226,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 1.9176266193389893,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209059233449477,
+ "grad_norm": 0.05939047038555145,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 1.9227800369262695,
+ "step": 2103
+ },
+ {
+ "epoch": 29.222996515679444,
+ "grad_norm": 0.059280600398778915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 1.9164855480194092,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23693379790941,
+ "grad_norm": 0.058746229857206345,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 1.9110995531082153,
+ "step": 2105
+ },
+ {
+ "epoch": 29.250871080139373,
+ "grad_norm": 0.05882417410612106,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 1.926957368850708,
+ "step": 2106
+ },
+ {
+ "epoch": 29.264808362369337,
+ "grad_norm": 0.05827188491821289,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 1.9101874828338623,
+ "step": 2107
+ },
+ {
+ "epoch": 29.278745644599304,
+ "grad_norm": 0.059739116579294205,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 1.924944519996643,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29268292682927,
+ "grad_norm": 0.058013755828142166,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 1.9121344089508057,
+ "step": 2109
+ },
+ {
+ "epoch": 29.306620209059233,
+ "grad_norm": 0.061207350343465805,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 1.9137616157531738,
+ "step": 2110
+ },
+ {
+ "epoch": 29.320557491289197,
+ "grad_norm": 0.05848725140094757,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 1.9290939569473267,
+ "step": 2111
+ },
+ {
+ "epoch": 29.334494773519165,
+ "grad_norm": 0.059134867042303085,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 1.919262409210205,
+ "step": 2112
+ },
+ {
+ "epoch": 29.34843205574913,
+ "grad_norm": 0.06025470420718193,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 1.9021251201629639,
+ "step": 2113
+ },
+ {
+ "epoch": 29.362369337979093,
+ "grad_norm": 0.05981536582112312,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 1.9163634777069092,
+ "step": 2114
+ },
+ {
+ "epoch": 29.37630662020906,
+ "grad_norm": 0.05826934054493904,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 1.8987491130828857,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390243902439025,
+ "grad_norm": 0.06095297262072563,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 1.9322285652160645,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40418118466899,
+ "grad_norm": 0.05814153701066971,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 1.927016019821167,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418118466898953,
+ "grad_norm": 0.05872226879000664,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 1.9271752834320068,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43205574912892,
+ "grad_norm": 0.06008293107151985,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 1.9269304275512695,
+ "step": 2119
+ },
+ {
+ "epoch": 29.445993031358885,
+ "grad_norm": 0.06035081297159195,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 1.929135799407959,
+ "step": 2120
+ },
+ {
+ "epoch": 29.45993031358885,
+ "grad_norm": 0.06080561876296997,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 1.9284312725067139,
+ "step": 2121
+ },
+ {
+ "epoch": 29.473867595818817,
+ "grad_norm": 0.058793697506189346,
+ "learning_rate": 0.000231465389734324,
+ "loss": 1.931719183921814,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48780487804878,
+ "grad_norm": 0.06385456770658493,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 1.9112051725387573,
+ "step": 2123
+ },
+ {
+ "epoch": 29.501742160278745,
+ "grad_norm": 0.06030426174402237,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 1.9254870414733887,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51567944250871,
+ "grad_norm": 0.059933315962553024,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 1.915220022201538,
+ "step": 2125
+ },
+ {
+ "epoch": 29.529616724738677,
+ "grad_norm": 0.06083494424819946,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 1.9391965866088867,
+ "step": 2126
+ },
+ {
+ "epoch": 29.54355400696864,
+ "grad_norm": 0.06049460917711258,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 1.9301774501800537,
+ "step": 2127
+ },
+ {
+ "epoch": 29.557491289198605,
+ "grad_norm": 0.05953316017985344,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 1.9150071144104004,
+ "step": 2128
+ },
+ {
+ "epoch": 29.571428571428573,
+ "grad_norm": 0.0611971914768219,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 1.9265213012695312,
+ "step": 2129
+ },
+ {
+ "epoch": 29.585365853658537,
+ "grad_norm": 0.06123765930533409,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 1.9405107498168945,
+ "step": 2130
+ },
+ {
+ "epoch": 29.5993031358885,
+ "grad_norm": 0.06403640657663345,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 1.9216606616973877,
+ "step": 2131
+ },
+ {
+ "epoch": 29.613240418118465,
+ "grad_norm": 0.06530711054801941,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 1.9302153587341309,
+ "step": 2132
+ },
+ {
+ "epoch": 29.627177700348433,
+ "grad_norm": 0.06106860563158989,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 1.930979609489441,
+ "step": 2133
+ },
+ {
+ "epoch": 29.641114982578397,
+ "grad_norm": 0.06148209050297737,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 1.9276363849639893,
+ "step": 2134
+ },
+ {
+ "epoch": 29.65505226480836,
+ "grad_norm": 0.060283321887254715,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 1.9269459247589111,
+ "step": 2135
+ },
+ {
+ "epoch": 29.66898954703833,
+ "grad_norm": 0.06138703227043152,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 1.9389945268630981,
+ "step": 2136
+ },
+ {
+ "epoch": 29.682926829268293,
+ "grad_norm": 0.06101245433092117,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 1.9426372051239014,
+ "step": 2137
+ },
+ {
+ "epoch": 29.696864111498257,
+ "grad_norm": 0.05958810821175575,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 1.9327162504196167,
+ "step": 2138
+ },
+ {
+ "epoch": 29.71080139372822,
+ "grad_norm": 0.06250876933336258,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 1.9384305477142334,
+ "step": 2139
+ },
+ {
+ "epoch": 29.72473867595819,
+ "grad_norm": 0.06344381719827652,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 1.9422640800476074,
+ "step": 2140
+ },
+ {
+ "epoch": 29.738675958188153,
+ "grad_norm": 0.06065250560641289,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 1.920068621635437,
+ "step": 2141
+ },
+ {
+ "epoch": 29.752613240418118,
+ "grad_norm": 0.06333960592746735,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 1.925317645072937,
+ "step": 2142
+ },
+ {
+ "epoch": 29.766550522648085,
+ "grad_norm": 0.06285370141267776,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 1.92912757396698,
+ "step": 2143
+ },
+ {
+ "epoch": 29.78048780487805,
+ "grad_norm": 0.06441807746887207,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 1.941272497177124,
+ "step": 2144
+ },
+ {
+ "epoch": 29.794425087108014,
+ "grad_norm": 0.06314355880022049,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 1.923574447631836,
+ "step": 2145
+ },
+ {
+ "epoch": 29.808362369337978,
+ "grad_norm": 0.0639389157295227,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 1.9339193105697632,
+ "step": 2146
+ },
+ {
+ "epoch": 29.822299651567945,
+ "grad_norm": 0.06419187039136887,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 1.9294497966766357,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83623693379791,
+ "grad_norm": 0.06351246684789658,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 1.9282407760620117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.850174216027874,
+ "grad_norm": 0.06371650099754333,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 1.914009690284729,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86411149825784,
+ "grad_norm": 0.06891132146120071,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 1.948460578918457,
+ "step": 2150
+ },
+ {
+ "epoch": 29.878048780487806,
+ "grad_norm": 0.06575452536344528,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 1.9238896369934082,
+ "step": 2151
+ },
+ {
+ "epoch": 29.89198606271777,
+ "grad_norm": 0.06279128044843674,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 1.9414188861846924,
+ "step": 2152
+ },
+ {
+ "epoch": 29.905923344947734,
+ "grad_norm": 0.06493876129388809,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 1.9337334632873535,
+ "step": 2153
+ },
+ {
+ "epoch": 29.9198606271777,
+ "grad_norm": 0.0640307068824768,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 1.9258478879928589,
+ "step": 2154
+ },
+ {
+ "epoch": 29.933797909407666,
+ "grad_norm": 0.06517527252435684,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 1.9236748218536377,
+ "step": 2155
+ },
+ {
+ "epoch": 29.94773519163763,
+ "grad_norm": 0.06373083591461182,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 1.9360520839691162,
+ "step": 2156
+ },
+ {
+ "epoch": 29.961672473867594,
+ "grad_norm": 0.06771238893270493,
+ "learning_rate": 0.00022244633283095,
+ "loss": 1.947780966758728,
+ "step": 2157
+ },
+ {
+ "epoch": 29.975609756097562,
+ "grad_norm": 0.06298559159040451,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 1.9223754405975342,
+ "step": 2158
+ },
+ {
+ "epoch": 29.989547038327526,
+ "grad_norm": 0.06455206871032715,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 1.9330079555511475,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.0522056445479393,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 1.4463932514190674,
+ "step": 2160
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.232968049282253e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2160/training_args.bin b/runs/mt5-ssl/checkpoint-2160/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2160/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-2520/chat_template.jinja b/runs/mt5-ssl/checkpoint-2520/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-2520/config.json b/runs/mt5-ssl/checkpoint-2520/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-2520/generation_config.json b/runs/mt5-ssl/checkpoint-2520/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-2520/model.safetensors b/runs/mt5-ssl/checkpoint-2520/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..84e8f62144f6de3b66b74a9963c0a36ded5da9af
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d42b824537dda9424f95102e91c6cf2576552bf4b2a4a2b16b9bb1654c199b03
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-2520/optimizer.pt b/runs/mt5-ssl/checkpoint-2520/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c06672913503ded13751b3eb3243b701fec9074b
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:67b9b4b3eb4209c2137afbbb94c410e5ec505e8f5bbaa4ab715728d898dfd7da
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-2520/rng_state_0.pth b/runs/mt5-ssl/checkpoint-2520/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f6a767b5a760715c982d5f3b7abf7f7ae7024858
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:77a1aaa8594be4aeac424d02802142f8409ed1e2897c1e46e3c37ed12b93e83c
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2520/rng_state_1.pth b/runs/mt5-ssl/checkpoint-2520/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b0b2ce7324fa0e3d93e84f806452f72a932cb665
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8be9b03fd0264df00f58b952952d0e7adf408ae412c760c45b6919b81dd07768
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2520/scheduler.pt b/runs/mt5-ssl/checkpoint-2520/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..809e415b980a88e74111733a8fbd497a2408afe5
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0f5902b2719418cfd49abef69bf26be14b2bfe0841a35d7789b90886cf8ba0fb
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-2520/tokenizer.json b/runs/mt5-ssl/checkpoint-2520/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-2520/tokenizer_config.json b/runs/mt5-ssl/checkpoint-2520/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2520/trainer_state.json b/runs/mt5-ssl/checkpoint-2520/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..e8e551c7196fa192e4f63d3af63c2eced52f7e1a
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/trainer_state.json
@@ -0,0 +1,17946 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 35.0,
+ "eval_steps": 500,
+ "global_step": 2520,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6287577152252197,
+ "eval_runtime": 39.8508,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013937282229964,
+ "grad_norm": 0.056210409849882126,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 1.9758458137512207,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027874564459932,
+ "grad_norm": 0.06387963145971298,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 1.9743283987045288,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041811846689896,
+ "grad_norm": 0.059931688010692596,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 1.983224630355835,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05574912891986,
+ "grad_norm": 0.059698957949876785,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 1.9638850688934326,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069686411149824,
+ "grad_norm": 0.057556964457035065,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 1.9888463020324707,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083623693379792,
+ "grad_norm": 0.06250642985105515,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 1.9776880741119385,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097560975609756,
+ "grad_norm": 0.06218443810939789,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 1.9904823303222656,
+ "step": 1807
+ },
+ {
+ "epoch": 25.11149825783972,
+ "grad_norm": 0.06147411838173866,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 1.9881136417388916,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125435540069688,
+ "grad_norm": 0.06080890819430351,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 1.9833264350891113,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139372822299652,
+ "grad_norm": 0.057465583086013794,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 1.9764835834503174,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153310104529616,
+ "grad_norm": 0.061488356441259384,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 1.9807426929473877,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16724738675958,
+ "grad_norm": 0.060737013816833496,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 1.9796113967895508,
+ "step": 1812
+ },
+ {
+ "epoch": 25.181184668989548,
+ "grad_norm": 0.060508858412504196,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 1.9862518310546875,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195121951219512,
+ "grad_norm": 0.06062212213873863,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 1.9855456352233887,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209059233449477,
+ "grad_norm": 0.06143258139491081,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 1.9763805866241455,
+ "step": 1815
+ },
+ {
+ "epoch": 25.222996515679444,
+ "grad_norm": 0.06446489691734314,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 1.9674923419952393,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23693379790941,
+ "grad_norm": 0.06179432198405266,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 1.9792906045913696,
+ "step": 1817
+ },
+ {
+ "epoch": 25.250871080139373,
+ "grad_norm": 0.062219955027103424,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 1.983303189277649,
+ "step": 1818
+ },
+ {
+ "epoch": 25.264808362369337,
+ "grad_norm": 0.060631707310676575,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 1.9859293699264526,
+ "step": 1819
+ },
+ {
+ "epoch": 25.278745644599304,
+ "grad_norm": 0.06329842656850815,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 1.988373041152954,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29268292682927,
+ "grad_norm": 0.06290270388126373,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 1.9817759990692139,
+ "step": 1821
+ },
+ {
+ "epoch": 25.306620209059233,
+ "grad_norm": 0.062234487384557724,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 1.9615445137023926,
+ "step": 1822
+ },
+ {
+ "epoch": 25.320557491289197,
+ "grad_norm": 0.06063779070973396,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 1.9736756086349487,
+ "step": 1823
+ },
+ {
+ "epoch": 25.334494773519165,
+ "grad_norm": 0.06245938315987587,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 1.9932905435562134,
+ "step": 1824
+ },
+ {
+ "epoch": 25.34843205574913,
+ "grad_norm": 0.06380622833967209,
+ "learning_rate": 0.000309930621834911,
+ "loss": 1.9695440530776978,
+ "step": 1825
+ },
+ {
+ "epoch": 25.362369337979093,
+ "grad_norm": 0.0626329705119133,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 1.9820970296859741,
+ "step": 1826
+ },
+ {
+ "epoch": 25.37630662020906,
+ "grad_norm": 0.060844022780656815,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 1.9926050901412964,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390243902439025,
+ "grad_norm": 0.06239933893084526,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 2.001865863800049,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40418118466899,
+ "grad_norm": 0.06491564959287643,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 1.9897377490997314,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418118466898953,
+ "grad_norm": 0.05946580320596695,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 1.973170280456543,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43205574912892,
+ "grad_norm": 0.05930647253990173,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 1.9949123859405518,
+ "step": 1831
+ },
+ {
+ "epoch": 25.445993031358885,
+ "grad_norm": 0.06269685924053192,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 1.9912991523742676,
+ "step": 1832
+ },
+ {
+ "epoch": 25.45993031358885,
+ "grad_norm": 0.0637395903468132,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 1.9869414567947388,
+ "step": 1833
+ },
+ {
+ "epoch": 25.473867595818817,
+ "grad_norm": 0.0628778263926506,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 1.9748289585113525,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48780487804878,
+ "grad_norm": 0.06064455583691597,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 2.008216381072998,
+ "step": 1835
+ },
+ {
+ "epoch": 25.501742160278745,
+ "grad_norm": 0.0655415952205658,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 1.9852811098098755,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51567944250871,
+ "grad_norm": 0.06189798563718796,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 1.983716368675232,
+ "step": 1837
+ },
+ {
+ "epoch": 25.529616724738677,
+ "grad_norm": 0.061454080045223236,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 2.013745069503784,
+ "step": 1838
+ },
+ {
+ "epoch": 25.54355400696864,
+ "grad_norm": 0.06712181121110916,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 1.9856178760528564,
+ "step": 1839
+ },
+ {
+ "epoch": 25.557491289198605,
+ "grad_norm": 0.06364461034536362,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 2.0034379959106445,
+ "step": 1840
+ },
+ {
+ "epoch": 25.571428571428573,
+ "grad_norm": 0.06669855862855911,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 1.9868637323379517,
+ "step": 1841
+ },
+ {
+ "epoch": 25.585365853658537,
+ "grad_norm": 0.06578591465950012,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 1.9998185634613037,
+ "step": 1842
+ },
+ {
+ "epoch": 25.5993031358885,
+ "grad_norm": 0.06305381655693054,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 2.000594139099121,
+ "step": 1843
+ },
+ {
+ "epoch": 25.613240418118465,
+ "grad_norm": 0.061369556933641434,
+ "learning_rate": 0.000304866093757771,
+ "loss": 1.9811584949493408,
+ "step": 1844
+ },
+ {
+ "epoch": 25.627177700348433,
+ "grad_norm": 0.062390975654125214,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 2.0020859241485596,
+ "step": 1845
+ },
+ {
+ "epoch": 25.641114982578397,
+ "grad_norm": 0.06477145850658417,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 2.004000186920166,
+ "step": 1846
+ },
+ {
+ "epoch": 25.65505226480836,
+ "grad_norm": 0.06288585811853409,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 1.9919252395629883,
+ "step": 1847
+ },
+ {
+ "epoch": 25.66898954703833,
+ "grad_norm": 0.05959830805659294,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 1.9860963821411133,
+ "step": 1848
+ },
+ {
+ "epoch": 25.682926829268293,
+ "grad_norm": 0.0596182718873024,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 2.0014610290527344,
+ "step": 1849
+ },
+ {
+ "epoch": 25.696864111498257,
+ "grad_norm": 0.06665527820587158,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 2.0072989463806152,
+ "step": 1850
+ },
+ {
+ "epoch": 25.71080139372822,
+ "grad_norm": 0.0631684735417366,
+ "learning_rate": 0.000303,
+ "loss": 2.0076632499694824,
+ "step": 1851
+ },
+ {
+ "epoch": 25.72473867595819,
+ "grad_norm": 0.06214272975921631,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 2.0143537521362305,
+ "step": 1852
+ },
+ {
+ "epoch": 25.738675958188153,
+ "grad_norm": 0.06422685086727142,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 1.9885808229446411,
+ "step": 1853
+ },
+ {
+ "epoch": 25.752613240418118,
+ "grad_norm": 0.06583613157272339,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 2.0042953491210938,
+ "step": 1854
+ },
+ {
+ "epoch": 25.766550522648085,
+ "grad_norm": 0.06265029311180115,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 2.003157138824463,
+ "step": 1855
+ },
+ {
+ "epoch": 25.78048780487805,
+ "grad_norm": 0.0612902007997036,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 1.9832231998443604,
+ "step": 1856
+ },
+ {
+ "epoch": 25.794425087108014,
+ "grad_norm": 0.0625016912817955,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 2.0012407302856445,
+ "step": 1857
+ },
+ {
+ "epoch": 25.808362369337978,
+ "grad_norm": 0.062357913702726364,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 2.010338306427002,
+ "step": 1858
+ },
+ {
+ "epoch": 25.822299651567945,
+ "grad_norm": 0.06349784880876541,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 2.0105648040771484,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83623693379791,
+ "grad_norm": 0.061822179704904556,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 1.9987019300460815,
+ "step": 1860
+ },
+ {
+ "epoch": 25.850174216027874,
+ "grad_norm": 0.06443580985069275,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 1.994943618774414,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86411149825784,
+ "grad_norm": 0.061674103140830994,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 2.0116522312164307,
+ "step": 1862
+ },
+ {
+ "epoch": 25.878048780487806,
+ "grad_norm": 0.061944279819726944,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 2.012296438217163,
+ "step": 1863
+ },
+ {
+ "epoch": 25.89198606271777,
+ "grad_norm": 0.06270956248044968,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 2.0057339668273926,
+ "step": 1864
+ },
+ {
+ "epoch": 25.905923344947734,
+ "grad_norm": 0.06427973508834839,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 2.0053863525390625,
+ "step": 1865
+ },
+ {
+ "epoch": 25.9198606271777,
+ "grad_norm": 0.0637369304895401,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 2.0130667686462402,
+ "step": 1866
+ },
+ {
+ "epoch": 25.933797909407666,
+ "grad_norm": 0.06291835755109787,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 2.0094404220581055,
+ "step": 1867
+ },
+ {
+ "epoch": 25.94773519163763,
+ "grad_norm": 0.06374962627887726,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 2.0074503421783447,
+ "step": 1868
+ },
+ {
+ "epoch": 25.961672473867594,
+ "grad_norm": 0.06330475956201553,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 1.9999325275421143,
+ "step": 1869
+ },
+ {
+ "epoch": 25.975609756097562,
+ "grad_norm": 0.06077918782830238,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 1.994345784187317,
+ "step": 1870
+ },
+ {
+ "epoch": 25.989547038327526,
+ "grad_norm": 0.06524437665939331,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 2.0210394859313965,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.05064176768064499,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 1.4825201034545898,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6298588514328003,
+ "eval_runtime": 43.1768,
+ "eval_samples_per_second": 56.558,
+ "eval_steps_per_second": 0.463,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013937282229964,
+ "grad_norm": 0.05861504375934601,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 1.9728282690048218,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027874564459932,
+ "grad_norm": 0.06309985369443893,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 1.9638245105743408,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041811846689896,
+ "grad_norm": 0.06438111513853073,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 1.9697308540344238,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05574912891986,
+ "grad_norm": 0.060332708060741425,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 1.9574062824249268,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069686411149824,
+ "grad_norm": 0.061463311314582825,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 1.9469952583312988,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083623693379792,
+ "grad_norm": 0.06337480992078781,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 1.9695435762405396,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097560975609756,
+ "grad_norm": 0.06303273886442184,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 1.9764964580535889,
+ "step": 1879
+ },
+ {
+ "epoch": 26.11149825783972,
+ "grad_norm": 0.06061594933271408,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 1.9758808612823486,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125435540069688,
+ "grad_norm": 0.061181873083114624,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 1.9607892036437988,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139372822299652,
+ "grad_norm": 0.06557024270296097,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 1.9752333164215088,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153310104529616,
+ "grad_norm": 0.06261707097291946,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 1.9634147882461548,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16724738675958,
+ "grad_norm": 0.05993996560573578,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 1.957287311553955,
+ "step": 1884
+ },
+ {
+ "epoch": 26.181184668989548,
+ "grad_norm": 0.0612766332924366,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 1.9562265872955322,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195121951219512,
+ "grad_norm": 0.05735815688967705,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 1.9577300548553467,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209059233449477,
+ "grad_norm": 0.06258611381053925,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 1.9661660194396973,
+ "step": 1887
+ },
+ {
+ "epoch": 26.222996515679444,
+ "grad_norm": 0.06298062205314636,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 1.9508070945739746,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23693379790941,
+ "grad_norm": 0.06346822530031204,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 1.9677343368530273,
+ "step": 1889
+ },
+ {
+ "epoch": 26.250871080139373,
+ "grad_norm": 0.05959425866603851,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 1.9574894905090332,
+ "step": 1890
+ },
+ {
+ "epoch": 26.264808362369337,
+ "grad_norm": 0.0610729418694973,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 1.9515372514724731,
+ "step": 1891
+ },
+ {
+ "epoch": 26.278745644599304,
+ "grad_norm": 0.06362602859735489,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 1.966064453125,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29268292682927,
+ "grad_norm": 0.06124011054635048,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 1.9654541015625,
+ "step": 1893
+ },
+ {
+ "epoch": 26.306620209059233,
+ "grad_norm": 0.05773317068815231,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 1.9610934257507324,
+ "step": 1894
+ },
+ {
+ "epoch": 26.320557491289197,
+ "grad_norm": 0.05896785855293274,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 1.965133547782898,
+ "step": 1895
+ },
+ {
+ "epoch": 26.334494773519165,
+ "grad_norm": 0.059504859149456024,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 1.9617180824279785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.34843205574913,
+ "grad_norm": 0.05804497376084328,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 1.9761323928833008,
+ "step": 1897
+ },
+ {
+ "epoch": 26.362369337979093,
+ "grad_norm": 0.058211296796798706,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 1.9805936813354492,
+ "step": 1898
+ },
+ {
+ "epoch": 26.37630662020906,
+ "grad_norm": 0.05976752191781998,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 1.9514508247375488,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390243902439025,
+ "grad_norm": 0.05824121832847595,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 1.97285795211792,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40418118466899,
+ "grad_norm": 0.05793534219264984,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 1.9608209133148193,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418118466898953,
+ "grad_norm": 0.05697625130414963,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 1.9787089824676514,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43205574912892,
+ "grad_norm": 0.05703677609562874,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 1.9610021114349365,
+ "step": 1903
+ },
+ {
+ "epoch": 26.445993031358885,
+ "grad_norm": 0.057380311191082,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 1.9764081239700317,
+ "step": 1904
+ },
+ {
+ "epoch": 26.45993031358885,
+ "grad_norm": 0.05846763774752617,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 1.9641292095184326,
+ "step": 1905
+ },
+ {
+ "epoch": 26.473867595818817,
+ "grad_norm": 0.05929148569703102,
+ "learning_rate": 0.000288343693342466,
+ "loss": 1.9803684949874878,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48780487804878,
+ "grad_norm": 0.058236971497535706,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 1.981757402420044,
+ "step": 1907
+ },
+ {
+ "epoch": 26.501742160278745,
+ "grad_norm": 0.05845103785395622,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 1.9736125469207764,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51567944250871,
+ "grad_norm": 0.06104041263461113,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 1.9698188304901123,
+ "step": 1909
+ },
+ {
+ "epoch": 26.529616724738677,
+ "grad_norm": 0.0628688707947731,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 1.9786568880081177,
+ "step": 1910
+ },
+ {
+ "epoch": 26.54355400696864,
+ "grad_norm": 0.06115978956222534,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 1.9744799137115479,
+ "step": 1911
+ },
+ {
+ "epoch": 26.557491289198605,
+ "grad_norm": 0.06233362853527069,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 1.9801995754241943,
+ "step": 1912
+ },
+ {
+ "epoch": 26.571428571428573,
+ "grad_norm": 0.06182500347495079,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 1.9888432025909424,
+ "step": 1913
+ },
+ {
+ "epoch": 26.585365853658537,
+ "grad_norm": 0.05962767079472542,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 1.9648957252502441,
+ "step": 1914
+ },
+ {
+ "epoch": 26.5993031358885,
+ "grad_norm": 0.05994252488017082,
+ "learning_rate": 0.000285947841605349,
+ "loss": 1.9709041118621826,
+ "step": 1915
+ },
+ {
+ "epoch": 26.613240418118465,
+ "grad_norm": 0.06045769527554512,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 1.994588017463684,
+ "step": 1916
+ },
+ {
+ "epoch": 26.627177700348433,
+ "grad_norm": 0.059465229511260986,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 1.9743266105651855,
+ "step": 1917
+ },
+ {
+ "epoch": 26.641114982578397,
+ "grad_norm": 0.06020190939307213,
+ "learning_rate": 0.000285149463934261,
+ "loss": 1.9780995845794678,
+ "step": 1918
+ },
+ {
+ "epoch": 26.65505226480836,
+ "grad_norm": 0.05829813703894615,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 1.9831392765045166,
+ "step": 1919
+ },
+ {
+ "epoch": 26.66898954703833,
+ "grad_norm": 0.05960174649953842,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 1.9799871444702148,
+ "step": 1920
+ },
+ {
+ "epoch": 26.682926829268293,
+ "grad_norm": 0.06133116036653519,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 1.985158920288086,
+ "step": 1921
+ },
+ {
+ "epoch": 26.696864111498257,
+ "grad_norm": 0.05947015434503555,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 1.9783923625946045,
+ "step": 1922
+ },
+ {
+ "epoch": 26.71080139372822,
+ "grad_norm": 0.061687882989645004,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 1.9685261249542236,
+ "step": 1923
+ },
+ {
+ "epoch": 26.72473867595819,
+ "grad_norm": 0.0613567978143692,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 1.9945828914642334,
+ "step": 1924
+ },
+ {
+ "epoch": 26.738675958188153,
+ "grad_norm": 0.05902605876326561,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 1.966120719909668,
+ "step": 1925
+ },
+ {
+ "epoch": 26.752613240418118,
+ "grad_norm": 0.06018664687871933,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 1.991615653038025,
+ "step": 1926
+ },
+ {
+ "epoch": 26.766550522648085,
+ "grad_norm": 0.061171986162662506,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 1.9877536296844482,
+ "step": 1927
+ },
+ {
+ "epoch": 26.78048780487805,
+ "grad_norm": 0.06263386458158493,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 1.9799100160598755,
+ "step": 1928
+ },
+ {
+ "epoch": 26.794425087108014,
+ "grad_norm": 0.06326548755168915,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 1.9805757999420166,
+ "step": 1929
+ },
+ {
+ "epoch": 26.808362369337978,
+ "grad_norm": 0.06226228550076485,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 1.9867990016937256,
+ "step": 1930
+ },
+ {
+ "epoch": 26.822299651567945,
+ "grad_norm": 0.06003205478191376,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 1.9632174968719482,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83623693379791,
+ "grad_norm": 0.06075739488005638,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 1.9638378620147705,
+ "step": 1932
+ },
+ {
+ "epoch": 26.850174216027874,
+ "grad_norm": 0.061726443469524384,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 1.9865589141845703,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86411149825784,
+ "grad_norm": 0.0621049702167511,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 2.007071018218994,
+ "step": 1934
+ },
+ {
+ "epoch": 26.878048780487806,
+ "grad_norm": 0.061793744564056396,
+ "learning_rate": 0.000280627938758204,
+ "loss": 1.988924264907837,
+ "step": 1935
+ },
+ {
+ "epoch": 26.89198606271777,
+ "grad_norm": 0.062355563044548035,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 1.9830653667449951,
+ "step": 1936
+ },
+ {
+ "epoch": 26.905923344947734,
+ "grad_norm": 0.060881514102220535,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 2.003551959991455,
+ "step": 1937
+ },
+ {
+ "epoch": 26.9198606271777,
+ "grad_norm": 0.061703141778707504,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 1.9811534881591797,
+ "step": 1938
+ },
+ {
+ "epoch": 26.933797909407666,
+ "grad_norm": 0.0634910985827446,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 1.9840846061706543,
+ "step": 1939
+ },
+ {
+ "epoch": 26.94773519163763,
+ "grad_norm": 0.060807306319475174,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 1.9879541397094727,
+ "step": 1940
+ },
+ {
+ "epoch": 26.961672473867594,
+ "grad_norm": 0.06126011535525322,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 1.993988037109375,
+ "step": 1941
+ },
+ {
+ "epoch": 26.975609756097562,
+ "grad_norm": 0.06394185870885849,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 1.983365535736084,
+ "step": 1942
+ },
+ {
+ "epoch": 26.989547038327526,
+ "grad_norm": 0.06164516508579254,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 1.989990234375,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.04997039958834648,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 1.4921388626098633,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6310494542121887,
+ "eval_runtime": 40.875,
+ "eval_samples_per_second": 59.743,
+ "eval_steps_per_second": 0.489,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013937282229964,
+ "grad_norm": 0.05894223228096962,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 1.935828447341919,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027874564459932,
+ "grad_norm": 0.06731119751930237,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 1.945769190788269,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041811846689896,
+ "grad_norm": 0.06787799298763275,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 1.944061040878296,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05574912891986,
+ "grad_norm": 0.05840576812624931,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 1.94246506690979,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069686411149824,
+ "grad_norm": 0.058151841163635254,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 1.940029263496399,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083623693379792,
+ "grad_norm": 0.06329379975795746,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 1.9523427486419678,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097560975609756,
+ "grad_norm": 0.06530784070491791,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 1.9352362155914307,
+ "step": 1951
+ },
+ {
+ "epoch": 27.11149825783972,
+ "grad_norm": 0.06273649632930756,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 1.946252703666687,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125435540069688,
+ "grad_norm": 0.05996445193886757,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 1.9437150955200195,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139372822299652,
+ "grad_norm": 0.059087123721838,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 1.936135172843933,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153310104529616,
+ "grad_norm": 0.06257616728544235,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 1.9452544450759888,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16724738675958,
+ "grad_norm": 0.06591545045375824,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 1.9334893226623535,
+ "step": 1956
+ },
+ {
+ "epoch": 27.181184668989548,
+ "grad_norm": 0.0595872700214386,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 1.9514997005462646,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195121951219512,
+ "grad_norm": 0.05855429172515869,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 1.955148458480835,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209059233449477,
+ "grad_norm": 0.06383495032787323,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 1.9607688188552856,
+ "step": 1959
+ },
+ {
+ "epoch": 27.222996515679444,
+ "grad_norm": 0.06486682593822479,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 1.941009283065796,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23693379790941,
+ "grad_norm": 0.06293730437755585,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 1.9550879001617432,
+ "step": 1961
+ },
+ {
+ "epoch": 27.250871080139373,
+ "grad_norm": 0.06104227155447006,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 1.9483157396316528,
+ "step": 1962
+ },
+ {
+ "epoch": 27.264808362369337,
+ "grad_norm": 0.06920639425516129,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 1.948688268661499,
+ "step": 1963
+ },
+ {
+ "epoch": 27.278745644599304,
+ "grad_norm": 0.06186635419726372,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 1.9512240886688232,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29268292682927,
+ "grad_norm": 0.06258274614810944,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 1.9538028240203857,
+ "step": 1965
+ },
+ {
+ "epoch": 27.306620209059233,
+ "grad_norm": 0.06546445935964584,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 1.9497663974761963,
+ "step": 1966
+ },
+ {
+ "epoch": 27.320557491289197,
+ "grad_norm": 0.06289096921682358,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 1.9706182479858398,
+ "step": 1967
+ },
+ {
+ "epoch": 27.334494773519165,
+ "grad_norm": 0.06412214040756226,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 1.9423331022262573,
+ "step": 1968
+ },
+ {
+ "epoch": 27.34843205574913,
+ "grad_norm": 0.062773697078228,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 1.9471882581710815,
+ "step": 1969
+ },
+ {
+ "epoch": 27.362369337979093,
+ "grad_norm": 0.060290850698947906,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 1.9732742309570312,
+ "step": 1970
+ },
+ {
+ "epoch": 27.37630662020906,
+ "grad_norm": 0.062357380986213684,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 1.957550048828125,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390243902439025,
+ "grad_norm": 0.06070727854967117,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 1.951438069343567,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40418118466899,
+ "grad_norm": 0.06380726397037506,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 1.9512609243392944,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418118466898953,
+ "grad_norm": 0.06135795637965202,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 1.9663273096084595,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43205574912892,
+ "grad_norm": 0.0627567246556282,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 1.964203119277954,
+ "step": 1975
+ },
+ {
+ "epoch": 27.445993031358885,
+ "grad_norm": 0.06235141307115555,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 1.960247278213501,
+ "step": 1976
+ },
+ {
+ "epoch": 27.45993031358885,
+ "grad_norm": 0.06187676638364792,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 1.9667054414749146,
+ "step": 1977
+ },
+ {
+ "epoch": 27.473867595818817,
+ "grad_norm": 0.06090894341468811,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 1.9695777893066406,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48780487804878,
+ "grad_norm": 0.06363680213689804,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 1.9529361724853516,
+ "step": 1979
+ },
+ {
+ "epoch": 27.501742160278745,
+ "grad_norm": 0.06053898110985756,
+ "learning_rate": 0.00026868712586269,
+ "loss": 1.9707324504852295,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51567944250871,
+ "grad_norm": 0.061363156884908676,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 1.955542802810669,
+ "step": 1981
+ },
+ {
+ "epoch": 27.529616724738677,
+ "grad_norm": 0.06158224865794182,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 1.9489967823028564,
+ "step": 1982
+ },
+ {
+ "epoch": 27.54355400696864,
+ "grad_norm": 0.05886901170015335,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 1.9705257415771484,
+ "step": 1983
+ },
+ {
+ "epoch": 27.557491289198605,
+ "grad_norm": 0.05890485271811485,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 1.9430618286132812,
+ "step": 1984
+ },
+ {
+ "epoch": 27.571428571428573,
+ "grad_norm": 0.060433268547058105,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 1.9634191989898682,
+ "step": 1985
+ },
+ {
+ "epoch": 27.585365853658537,
+ "grad_norm": 0.0626354068517685,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 1.9568289518356323,
+ "step": 1986
+ },
+ {
+ "epoch": 27.5993031358885,
+ "grad_norm": 0.05973317474126816,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 1.9678292274475098,
+ "step": 1987
+ },
+ {
+ "epoch": 27.613240418118465,
+ "grad_norm": 0.062289632856845856,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 1.9413363933563232,
+ "step": 1988
+ },
+ {
+ "epoch": 27.627177700348433,
+ "grad_norm": 0.06162180379033089,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 1.9835751056671143,
+ "step": 1989
+ },
+ {
+ "epoch": 27.641114982578397,
+ "grad_norm": 0.06176188960671425,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 1.9541220664978027,
+ "step": 1990
+ },
+ {
+ "epoch": 27.65505226480836,
+ "grad_norm": 0.06479815393686295,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 1.9596972465515137,
+ "step": 1991
+ },
+ {
+ "epoch": 27.66898954703833,
+ "grad_norm": 0.06571854650974274,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 1.9628980159759521,
+ "step": 1992
+ },
+ {
+ "epoch": 27.682926829268293,
+ "grad_norm": 0.06167568638920784,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 1.9624769687652588,
+ "step": 1993
+ },
+ {
+ "epoch": 27.696864111498257,
+ "grad_norm": 0.06207694113254547,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 1.9487217664718628,
+ "step": 1994
+ },
+ {
+ "epoch": 27.71080139372822,
+ "grad_norm": 0.061430424451828,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 1.9567784070968628,
+ "step": 1995
+ },
+ {
+ "epoch": 27.72473867595819,
+ "grad_norm": 0.06589192897081375,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 1.9594354629516602,
+ "step": 1996
+ },
+ {
+ "epoch": 27.738675958188153,
+ "grad_norm": 0.06640033423900604,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 1.9637835025787354,
+ "step": 1997
+ },
+ {
+ "epoch": 27.752613240418118,
+ "grad_norm": 0.06181980296969414,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 1.9597129821777344,
+ "step": 1998
+ },
+ {
+ "epoch": 27.766550522648085,
+ "grad_norm": 0.06352445483207703,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 1.9612562656402588,
+ "step": 1999
+ },
+ {
+ "epoch": 27.78048780487805,
+ "grad_norm": 0.06398574262857437,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 1.9662930965423584,
+ "step": 2000
+ },
+ {
+ "epoch": 27.794425087108014,
+ "grad_norm": 0.06195084750652313,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 1.956035852432251,
+ "step": 2001
+ },
+ {
+ "epoch": 27.808362369337978,
+ "grad_norm": 0.05951351299881935,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 1.9712438583374023,
+ "step": 2002
+ },
+ {
+ "epoch": 27.822299651567945,
+ "grad_norm": 0.062185633927583694,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 1.9714539051055908,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83623693379791,
+ "grad_norm": 0.061178892850875854,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 1.9691438674926758,
+ "step": 2004
+ },
+ {
+ "epoch": 27.850174216027874,
+ "grad_norm": 0.06262431293725967,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 1.9506231546401978,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86411149825784,
+ "grad_norm": 0.06206081807613373,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 1.9662809371948242,
+ "step": 2006
+ },
+ {
+ "epoch": 27.878048780487806,
+ "grad_norm": 0.0626990869641304,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 1.972696304321289,
+ "step": 2007
+ },
+ {
+ "epoch": 27.89198606271777,
+ "grad_norm": 0.06345485150814056,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 1.9560344219207764,
+ "step": 2008
+ },
+ {
+ "epoch": 27.905923344947734,
+ "grad_norm": 0.06401040405035019,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 1.9732205867767334,
+ "step": 2009
+ },
+ {
+ "epoch": 27.9198606271777,
+ "grad_norm": 0.06437943130731583,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 1.974897027015686,
+ "step": 2010
+ },
+ {
+ "epoch": 27.933797909407666,
+ "grad_norm": 0.06296383589506149,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 1.9710102081298828,
+ "step": 2011
+ },
+ {
+ "epoch": 27.94773519163763,
+ "grad_norm": 0.06527025997638702,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 1.9707660675048828,
+ "step": 2012
+ },
+ {
+ "epoch": 27.961672473867594,
+ "grad_norm": 0.06185369938611984,
+ "learning_rate": 0.00025996500713765,
+ "loss": 1.9675114154815674,
+ "step": 2013
+ },
+ {
+ "epoch": 27.975609756097562,
+ "grad_norm": 0.06357269734144211,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 1.9812211990356445,
+ "step": 2014
+ },
+ {
+ "epoch": 27.989547038327526,
+ "grad_norm": 0.06581240892410278,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 1.97153639793396,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.049184102565050125,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 1.4647765159606934,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6328383088111877,
+ "eval_runtime": 40.5927,
+ "eval_samples_per_second": 60.159,
+ "eval_steps_per_second": 0.493,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013937282229964,
+ "grad_norm": 0.05999046564102173,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 1.9182333946228027,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027874564459932,
+ "grad_norm": 0.059804853051900864,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 1.924762487411499,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041811846689896,
+ "grad_norm": 0.05890650674700737,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 1.9341628551483154,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05574912891986,
+ "grad_norm": 0.057233694940805435,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 1.932476282119751,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069686411149824,
+ "grad_norm": 0.060101717710494995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 1.9403120279312134,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083623693379792,
+ "grad_norm": 0.06054912135004997,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 1.9274934530258179,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097560975609756,
+ "grad_norm": 0.05882031098008156,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 1.9331915378570557,
+ "step": 2023
+ },
+ {
+ "epoch": 28.11149825783972,
+ "grad_norm": 0.05846688151359558,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 1.9342529773712158,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125435540069688,
+ "grad_norm": 0.058974213898181915,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 1.9353152513504028,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139372822299652,
+ "grad_norm": 0.05882132798433304,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 1.9500072002410889,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153310104529616,
+ "grad_norm": 0.05919627472758293,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 1.9313960075378418,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16724738675958,
+ "grad_norm": 0.06413958966732025,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 1.9516873359680176,
+ "step": 2028
+ },
+ {
+ "epoch": 28.181184668989548,
+ "grad_norm": 0.05799189209938049,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 1.9290452003479004,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195121951219512,
+ "grad_norm": 0.05999152734875679,
+ "learning_rate": 0.000255486047794226,
+ "loss": 1.9396628141403198,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209059233449477,
+ "grad_norm": 0.060240138322114944,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 1.937988519668579,
+ "step": 2031
+ },
+ {
+ "epoch": 28.222996515679444,
+ "grad_norm": 0.05933740735054016,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 1.9343246221542358,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23693379790941,
+ "grad_norm": 0.05960093438625336,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 1.9326903820037842,
+ "step": 2033
+ },
+ {
+ "epoch": 28.250871080139373,
+ "grad_norm": 0.0635334923863411,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 1.9381543397903442,
+ "step": 2034
+ },
+ {
+ "epoch": 28.264808362369337,
+ "grad_norm": 0.06162877380847931,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 1.9265705347061157,
+ "step": 2035
+ },
+ {
+ "epoch": 28.278745644599304,
+ "grad_norm": 0.059690047055482864,
+ "learning_rate": 0.000253907826333243,
+ "loss": 1.9278099536895752,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29268292682927,
+ "grad_norm": 0.0612013153731823,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 1.939026117324829,
+ "step": 2037
+ },
+ {
+ "epoch": 28.306620209059233,
+ "grad_norm": 0.06080419942736626,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 1.9306461811065674,
+ "step": 2038
+ },
+ {
+ "epoch": 28.320557491289197,
+ "grad_norm": 0.06308622658252716,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 1.9350788593292236,
+ "step": 2039
+ },
+ {
+ "epoch": 28.334494773519165,
+ "grad_norm": 0.06124545633792877,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 1.9264733791351318,
+ "step": 2040
+ },
+ {
+ "epoch": 28.34843205574913,
+ "grad_norm": 0.059037938714027405,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 1.917725920677185,
+ "step": 2041
+ },
+ {
+ "epoch": 28.362369337979093,
+ "grad_norm": 0.06089222431182861,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 1.949327826499939,
+ "step": 2042
+ },
+ {
+ "epoch": 28.37630662020906,
+ "grad_norm": 0.05957084521651268,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 1.938978910446167,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390243902439025,
+ "grad_norm": 0.06060048192739487,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 1.9226362705230713,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40418118466899,
+ "grad_norm": 0.058304425328969955,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 1.9330644607543945,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418118466898953,
+ "grad_norm": 0.05970550701022148,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 1.931627631187439,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43205574912892,
+ "grad_norm": 0.05994785949587822,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 1.9387202262878418,
+ "step": 2047
+ },
+ {
+ "epoch": 28.445993031358885,
+ "grad_norm": 0.06001518666744232,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 1.951300859451294,
+ "step": 2048
+ },
+ {
+ "epoch": 28.45993031358885,
+ "grad_norm": 0.0636107549071312,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 1.9358865022659302,
+ "step": 2049
+ },
+ {
+ "epoch": 28.473867595818817,
+ "grad_norm": 0.059402115643024445,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 1.9298653602600098,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48780487804878,
+ "grad_norm": 0.06260179728269577,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 1.954633116722107,
+ "step": 2051
+ },
+ {
+ "epoch": 28.501742160278745,
+ "grad_norm": 0.06439191848039627,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 1.9338306188583374,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51567944250871,
+ "grad_norm": 0.06265310943126678,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 1.9500043392181396,
+ "step": 2053
+ },
+ {
+ "epoch": 28.529616724738677,
+ "grad_norm": 0.06323659420013428,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 1.9445586204528809,
+ "step": 2054
+ },
+ {
+ "epoch": 28.54355400696864,
+ "grad_norm": 0.06512445956468582,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 1.958587408065796,
+ "step": 2055
+ },
+ {
+ "epoch": 28.557491289198605,
+ "grad_norm": 0.06386571377515793,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 1.949887990951538,
+ "step": 2056
+ },
+ {
+ "epoch": 28.571428571428573,
+ "grad_norm": 0.06317705661058426,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 1.9512674808502197,
+ "step": 2057
+ },
+ {
+ "epoch": 28.585365853658537,
+ "grad_norm": 0.06504984200000763,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 1.9242960214614868,
+ "step": 2058
+ },
+ {
+ "epoch": 28.5993031358885,
+ "grad_norm": 0.06278593838214874,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 1.9203131198883057,
+ "step": 2059
+ },
+ {
+ "epoch": 28.613240418118465,
+ "grad_norm": 0.06171257048845291,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 1.954156517982483,
+ "step": 2060
+ },
+ {
+ "epoch": 28.627177700348433,
+ "grad_norm": 0.06303241103887558,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 1.9460458755493164,
+ "step": 2061
+ },
+ {
+ "epoch": 28.641114982578397,
+ "grad_norm": 0.06252635270357132,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 1.9477293491363525,
+ "step": 2062
+ },
+ {
+ "epoch": 28.65505226480836,
+ "grad_norm": 0.06226412579417229,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 1.9259647130966187,
+ "step": 2063
+ },
+ {
+ "epoch": 28.66898954703833,
+ "grad_norm": 0.06111626327037811,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 1.9573581218719482,
+ "step": 2064
+ },
+ {
+ "epoch": 28.682926829268293,
+ "grad_norm": 0.06460461020469666,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 1.9486393928527832,
+ "step": 2065
+ },
+ {
+ "epoch": 28.696864111498257,
+ "grad_norm": 0.062027834355831146,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 1.9404425621032715,
+ "step": 2066
+ },
+ {
+ "epoch": 28.71080139372822,
+ "grad_norm": 0.061942074447870255,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 1.9444060325622559,
+ "step": 2067
+ },
+ {
+ "epoch": 28.72473867595819,
+ "grad_norm": 0.06159934028983116,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 1.9635517597198486,
+ "step": 2068
+ },
+ {
+ "epoch": 28.738675958188153,
+ "grad_norm": 0.0621839202940464,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 1.9442765712738037,
+ "step": 2069
+ },
+ {
+ "epoch": 28.752613240418118,
+ "grad_norm": 0.06432677060365677,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 1.9604840278625488,
+ "step": 2070
+ },
+ {
+ "epoch": 28.766550522648085,
+ "grad_norm": 0.06037316843867302,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 1.948632836341858,
+ "step": 2071
+ },
+ {
+ "epoch": 28.78048780487805,
+ "grad_norm": 0.05999939888715744,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 1.9451751708984375,
+ "step": 2072
+ },
+ {
+ "epoch": 28.794425087108014,
+ "grad_norm": 0.06441698968410492,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 1.9356517791748047,
+ "step": 2073
+ },
+ {
+ "epoch": 28.808362369337978,
+ "grad_norm": 0.06059470772743225,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 1.9258331060409546,
+ "step": 2074
+ },
+ {
+ "epoch": 28.822299651567945,
+ "grad_norm": 0.062317874282598495,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 1.9342000484466553,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83623693379791,
+ "grad_norm": 0.06318258494138718,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 1.9395548105239868,
+ "step": 2076
+ },
+ {
+ "epoch": 28.850174216027874,
+ "grad_norm": 0.06204817444086075,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 1.9572374820709229,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86411149825784,
+ "grad_norm": 0.06250188499689102,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 1.9465739727020264,
+ "step": 2078
+ },
+ {
+ "epoch": 28.878048780487806,
+ "grad_norm": 0.06293781846761703,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 1.9431169033050537,
+ "step": 2079
+ },
+ {
+ "epoch": 28.89198606271777,
+ "grad_norm": 0.059349458664655685,
+ "learning_rate": 0.000242380656502223,
+ "loss": 1.9673197269439697,
+ "step": 2080
+ },
+ {
+ "epoch": 28.905923344947734,
+ "grad_norm": 0.06214660778641701,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 1.9646787643432617,
+ "step": 2081
+ },
+ {
+ "epoch": 28.9198606271777,
+ "grad_norm": 0.062279123812913895,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 1.9572803974151611,
+ "step": 2082
+ },
+ {
+ "epoch": 28.933797909407666,
+ "grad_norm": 0.06502588093280792,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 1.9510462284088135,
+ "step": 2083
+ },
+ {
+ "epoch": 28.94773519163763,
+ "grad_norm": 0.06153101101517677,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 1.9507923126220703,
+ "step": 2084
+ },
+ {
+ "epoch": 28.961672473867594,
+ "grad_norm": 0.06469229608774185,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 1.937751293182373,
+ "step": 2085
+ },
+ {
+ "epoch": 28.975609756097562,
+ "grad_norm": 0.0672568678855896,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 1.944124698638916,
+ "step": 2086
+ },
+ {
+ "epoch": 28.989547038327526,
+ "grad_norm": 0.06451335549354553,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 1.9519702196121216,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.05141456052660942,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 1.4445865154266357,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6343450546264648,
+ "eval_runtime": 40.5081,
+ "eval_samples_per_second": 60.284,
+ "eval_steps_per_second": 0.494,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013937282229964,
+ "grad_norm": 0.061562713235616684,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 1.921946406364441,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027874564459932,
+ "grad_norm": 0.06396839767694473,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 1.9222416877746582,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041811846689896,
+ "grad_norm": 0.05828692391514778,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 1.9173905849456787,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05574912891986,
+ "grad_norm": 0.06567556411027908,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 1.9148579835891724,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069686411149824,
+ "grad_norm": 0.05924173817038536,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 1.916333794593811,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083623693379792,
+ "grad_norm": 0.06129243224859238,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 1.9031038284301758,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097560975609756,
+ "grad_norm": 0.06068028509616852,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 1.9165983200073242,
+ "step": 2095
+ },
+ {
+ "epoch": 29.11149825783972,
+ "grad_norm": 0.06304074823856354,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 1.90432870388031,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125435540069688,
+ "grad_norm": 0.060179345309734344,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 1.9187151193618774,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139372822299652,
+ "grad_norm": 0.05633116513490677,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 1.9125087261199951,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153310104529616,
+ "grad_norm": 0.0598532073199749,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 1.906031847000122,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16724738675958,
+ "grad_norm": 0.05987119674682617,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 1.9177607297897339,
+ "step": 2100
+ },
+ {
+ "epoch": 29.181184668989548,
+ "grad_norm": 0.060883522033691406,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 1.9008433818817139,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195121951219512,
+ "grad_norm": 0.05977628752589226,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 1.9176266193389893,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209059233449477,
+ "grad_norm": 0.05939047038555145,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 1.9227800369262695,
+ "step": 2103
+ },
+ {
+ "epoch": 29.222996515679444,
+ "grad_norm": 0.059280600398778915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 1.9164855480194092,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23693379790941,
+ "grad_norm": 0.058746229857206345,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 1.9110995531082153,
+ "step": 2105
+ },
+ {
+ "epoch": 29.250871080139373,
+ "grad_norm": 0.05882417410612106,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 1.926957368850708,
+ "step": 2106
+ },
+ {
+ "epoch": 29.264808362369337,
+ "grad_norm": 0.05827188491821289,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 1.9101874828338623,
+ "step": 2107
+ },
+ {
+ "epoch": 29.278745644599304,
+ "grad_norm": 0.059739116579294205,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 1.924944519996643,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29268292682927,
+ "grad_norm": 0.058013755828142166,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 1.9121344089508057,
+ "step": 2109
+ },
+ {
+ "epoch": 29.306620209059233,
+ "grad_norm": 0.061207350343465805,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 1.9137616157531738,
+ "step": 2110
+ },
+ {
+ "epoch": 29.320557491289197,
+ "grad_norm": 0.05848725140094757,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 1.9290939569473267,
+ "step": 2111
+ },
+ {
+ "epoch": 29.334494773519165,
+ "grad_norm": 0.059134867042303085,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 1.919262409210205,
+ "step": 2112
+ },
+ {
+ "epoch": 29.34843205574913,
+ "grad_norm": 0.06025470420718193,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 1.9021251201629639,
+ "step": 2113
+ },
+ {
+ "epoch": 29.362369337979093,
+ "grad_norm": 0.05981536582112312,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 1.9163634777069092,
+ "step": 2114
+ },
+ {
+ "epoch": 29.37630662020906,
+ "grad_norm": 0.05826934054493904,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 1.8987491130828857,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390243902439025,
+ "grad_norm": 0.06095297262072563,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 1.9322285652160645,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40418118466899,
+ "grad_norm": 0.05814153701066971,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 1.927016019821167,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418118466898953,
+ "grad_norm": 0.05872226879000664,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 1.9271752834320068,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43205574912892,
+ "grad_norm": 0.06008293107151985,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 1.9269304275512695,
+ "step": 2119
+ },
+ {
+ "epoch": 29.445993031358885,
+ "grad_norm": 0.06035081297159195,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 1.929135799407959,
+ "step": 2120
+ },
+ {
+ "epoch": 29.45993031358885,
+ "grad_norm": 0.06080561876296997,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 1.9284312725067139,
+ "step": 2121
+ },
+ {
+ "epoch": 29.473867595818817,
+ "grad_norm": 0.058793697506189346,
+ "learning_rate": 0.000231465389734324,
+ "loss": 1.931719183921814,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48780487804878,
+ "grad_norm": 0.06385456770658493,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 1.9112051725387573,
+ "step": 2123
+ },
+ {
+ "epoch": 29.501742160278745,
+ "grad_norm": 0.06030426174402237,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 1.9254870414733887,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51567944250871,
+ "grad_norm": 0.059933315962553024,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 1.915220022201538,
+ "step": 2125
+ },
+ {
+ "epoch": 29.529616724738677,
+ "grad_norm": 0.06083494424819946,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 1.9391965866088867,
+ "step": 2126
+ },
+ {
+ "epoch": 29.54355400696864,
+ "grad_norm": 0.06049460917711258,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 1.9301774501800537,
+ "step": 2127
+ },
+ {
+ "epoch": 29.557491289198605,
+ "grad_norm": 0.05953316017985344,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 1.9150071144104004,
+ "step": 2128
+ },
+ {
+ "epoch": 29.571428571428573,
+ "grad_norm": 0.0611971914768219,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 1.9265213012695312,
+ "step": 2129
+ },
+ {
+ "epoch": 29.585365853658537,
+ "grad_norm": 0.06123765930533409,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 1.9405107498168945,
+ "step": 2130
+ },
+ {
+ "epoch": 29.5993031358885,
+ "grad_norm": 0.06403640657663345,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 1.9216606616973877,
+ "step": 2131
+ },
+ {
+ "epoch": 29.613240418118465,
+ "grad_norm": 0.06530711054801941,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 1.9302153587341309,
+ "step": 2132
+ },
+ {
+ "epoch": 29.627177700348433,
+ "grad_norm": 0.06106860563158989,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 1.930979609489441,
+ "step": 2133
+ },
+ {
+ "epoch": 29.641114982578397,
+ "grad_norm": 0.06148209050297737,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 1.9276363849639893,
+ "step": 2134
+ },
+ {
+ "epoch": 29.65505226480836,
+ "grad_norm": 0.060283321887254715,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 1.9269459247589111,
+ "step": 2135
+ },
+ {
+ "epoch": 29.66898954703833,
+ "grad_norm": 0.06138703227043152,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 1.9389945268630981,
+ "step": 2136
+ },
+ {
+ "epoch": 29.682926829268293,
+ "grad_norm": 0.06101245433092117,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 1.9426372051239014,
+ "step": 2137
+ },
+ {
+ "epoch": 29.696864111498257,
+ "grad_norm": 0.05958810821175575,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 1.9327162504196167,
+ "step": 2138
+ },
+ {
+ "epoch": 29.71080139372822,
+ "grad_norm": 0.06250876933336258,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 1.9384305477142334,
+ "step": 2139
+ },
+ {
+ "epoch": 29.72473867595819,
+ "grad_norm": 0.06344381719827652,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 1.9422640800476074,
+ "step": 2140
+ },
+ {
+ "epoch": 29.738675958188153,
+ "grad_norm": 0.06065250560641289,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 1.920068621635437,
+ "step": 2141
+ },
+ {
+ "epoch": 29.752613240418118,
+ "grad_norm": 0.06333960592746735,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 1.925317645072937,
+ "step": 2142
+ },
+ {
+ "epoch": 29.766550522648085,
+ "grad_norm": 0.06285370141267776,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 1.92912757396698,
+ "step": 2143
+ },
+ {
+ "epoch": 29.78048780487805,
+ "grad_norm": 0.06441807746887207,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 1.941272497177124,
+ "step": 2144
+ },
+ {
+ "epoch": 29.794425087108014,
+ "grad_norm": 0.06314355880022049,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 1.923574447631836,
+ "step": 2145
+ },
+ {
+ "epoch": 29.808362369337978,
+ "grad_norm": 0.0639389157295227,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 1.9339193105697632,
+ "step": 2146
+ },
+ {
+ "epoch": 29.822299651567945,
+ "grad_norm": 0.06419187039136887,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 1.9294497966766357,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83623693379791,
+ "grad_norm": 0.06351246684789658,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 1.9282407760620117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.850174216027874,
+ "grad_norm": 0.06371650099754333,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 1.914009690284729,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86411149825784,
+ "grad_norm": 0.06891132146120071,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 1.948460578918457,
+ "step": 2150
+ },
+ {
+ "epoch": 29.878048780487806,
+ "grad_norm": 0.06575452536344528,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 1.9238896369934082,
+ "step": 2151
+ },
+ {
+ "epoch": 29.89198606271777,
+ "grad_norm": 0.06279128044843674,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 1.9414188861846924,
+ "step": 2152
+ },
+ {
+ "epoch": 29.905923344947734,
+ "grad_norm": 0.06493876129388809,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 1.9337334632873535,
+ "step": 2153
+ },
+ {
+ "epoch": 29.9198606271777,
+ "grad_norm": 0.0640307068824768,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 1.9258478879928589,
+ "step": 2154
+ },
+ {
+ "epoch": 29.933797909407666,
+ "grad_norm": 0.06517527252435684,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 1.9236748218536377,
+ "step": 2155
+ },
+ {
+ "epoch": 29.94773519163763,
+ "grad_norm": 0.06373083591461182,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 1.9360520839691162,
+ "step": 2156
+ },
+ {
+ "epoch": 29.961672473867594,
+ "grad_norm": 0.06771238893270493,
+ "learning_rate": 0.00022244633283095,
+ "loss": 1.947780966758728,
+ "step": 2157
+ },
+ {
+ "epoch": 29.975609756097562,
+ "grad_norm": 0.06298559159040451,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 1.9223754405975342,
+ "step": 2158
+ },
+ {
+ "epoch": 29.989547038327526,
+ "grad_norm": 0.06455206871032715,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 1.9330079555511475,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.0522056445479393,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 1.4463932514190674,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6358165740966797,
+ "eval_runtime": 40.2782,
+ "eval_samples_per_second": 60.628,
+ "eval_steps_per_second": 0.497,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013937282229964,
+ "grad_norm": 0.06026934087276459,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 1.9035022258758545,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027874564459932,
+ "grad_norm": 0.06600011885166168,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 1.8965950012207031,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041811846689896,
+ "grad_norm": 0.060655586421489716,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 1.8954249620437622,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05574912891986,
+ "grad_norm": 0.0636422410607338,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 1.8951303958892822,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069686411149824,
+ "grad_norm": 0.05972317233681679,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 1.910957932472229,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083623693379792,
+ "grad_norm": 0.060592107474803925,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 1.9128470420837402,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097560975609756,
+ "grad_norm": 0.05888872966170311,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 1.897517204284668,
+ "step": 2167
+ },
+ {
+ "epoch": 30.11149825783972,
+ "grad_norm": 0.062287602573633194,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 1.9048705101013184,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125435540069688,
+ "grad_norm": 0.05919402465224266,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 1.901914119720459,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139372822299652,
+ "grad_norm": 0.06069913133978844,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 1.9066851139068604,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153310104529616,
+ "grad_norm": 0.06046717241406441,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 1.8969058990478516,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16724738675958,
+ "grad_norm": 0.059624481946229935,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 1.8908498287200928,
+ "step": 2172
+ },
+ {
+ "epoch": 30.181184668989548,
+ "grad_norm": 0.06100156158208847,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 1.909839391708374,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195121951219512,
+ "grad_norm": 0.05801135674118996,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 1.9057464599609375,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209059233449477,
+ "grad_norm": 0.06198010593652725,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 1.9051663875579834,
+ "step": 2175
+ },
+ {
+ "epoch": 30.222996515679444,
+ "grad_norm": 0.06107710674405098,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 1.904172658920288,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23693379790941,
+ "grad_norm": 0.060085635632276535,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 1.8970673084259033,
+ "step": 2177
+ },
+ {
+ "epoch": 30.250871080139373,
+ "grad_norm": 0.05755467712879181,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 1.9092798233032227,
+ "step": 2178
+ },
+ {
+ "epoch": 30.264808362369337,
+ "grad_norm": 0.05827959626913071,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 1.9012596607208252,
+ "step": 2179
+ },
+ {
+ "epoch": 30.278745644599304,
+ "grad_norm": 0.0570731982588768,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 1.9027369022369385,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29268292682927,
+ "grad_norm": 0.05898377671837807,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 1.9284251928329468,
+ "step": 2181
+ },
+ {
+ "epoch": 30.306620209059233,
+ "grad_norm": 0.06119367107748985,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 1.8992201089859009,
+ "step": 2182
+ },
+ {
+ "epoch": 30.320557491289197,
+ "grad_norm": 0.05790785327553749,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 1.9106863737106323,
+ "step": 2183
+ },
+ {
+ "epoch": 30.334494773519165,
+ "grad_norm": 0.06101157143712044,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 1.9129860401153564,
+ "step": 2184
+ },
+ {
+ "epoch": 30.34843205574913,
+ "grad_norm": 0.062228068709373474,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 1.9075040817260742,
+ "step": 2185
+ },
+ {
+ "epoch": 30.362369337979093,
+ "grad_norm": 0.057724758982658386,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 1.9021648168563843,
+ "step": 2186
+ },
+ {
+ "epoch": 30.37630662020906,
+ "grad_norm": 0.05886731296777725,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 1.8955028057098389,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390243902439025,
+ "grad_norm": 0.0613555908203125,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 1.8902294635772705,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40418118466899,
+ "grad_norm": 0.059424031525850296,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 1.9234192371368408,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418118466898953,
+ "grad_norm": 0.059314578771591187,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 1.9120237827301025,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43205574912892,
+ "grad_norm": 0.059916652739048004,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 1.9118289947509766,
+ "step": 2191
+ },
+ {
+ "epoch": 30.445993031358885,
+ "grad_norm": 0.05817565694451332,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 1.9012054204940796,
+ "step": 2192
+ },
+ {
+ "epoch": 30.45993031358885,
+ "grad_norm": 0.061957065016031265,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 1.9110759496688843,
+ "step": 2193
+ },
+ {
+ "epoch": 30.473867595818817,
+ "grad_norm": 0.05899856239557266,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 1.9285509586334229,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48780487804878,
+ "grad_norm": 0.06001519039273262,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 1.9093801975250244,
+ "step": 2195
+ },
+ {
+ "epoch": 30.501742160278745,
+ "grad_norm": 0.060825154185295105,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 1.892765760421753,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51567944250871,
+ "grad_norm": 0.06136799231171608,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 1.9276010990142822,
+ "step": 2197
+ },
+ {
+ "epoch": 30.529616724738677,
+ "grad_norm": 0.05773079767823219,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 1.9083237648010254,
+ "step": 2198
+ },
+ {
+ "epoch": 30.54355400696864,
+ "grad_norm": 0.06120939180254936,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 1.8891932964324951,
+ "step": 2199
+ },
+ {
+ "epoch": 30.557491289198605,
+ "grad_norm": 0.06097486987709999,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 1.9084399938583374,
+ "step": 2200
+ },
+ {
+ "epoch": 30.571428571428573,
+ "grad_norm": 0.05973029136657715,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 1.918110966682434,
+ "step": 2201
+ },
+ {
+ "epoch": 30.585365853658537,
+ "grad_norm": 0.06335631012916565,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 1.9174532890319824,
+ "step": 2202
+ },
+ {
+ "epoch": 30.5993031358885,
+ "grad_norm": 0.05974910035729408,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 1.9107253551483154,
+ "step": 2203
+ },
+ {
+ "epoch": 30.613240418118465,
+ "grad_norm": 0.06281539052724838,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 1.9007208347320557,
+ "step": 2204
+ },
+ {
+ "epoch": 30.627177700348433,
+ "grad_norm": 0.062120143324136734,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 1.911742925643921,
+ "step": 2205
+ },
+ {
+ "epoch": 30.641114982578397,
+ "grad_norm": 0.06305570155382156,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 1.9121983051300049,
+ "step": 2206
+ },
+ {
+ "epoch": 30.65505226480836,
+ "grad_norm": 0.06112781539559364,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 1.908568263053894,
+ "step": 2207
+ },
+ {
+ "epoch": 30.66898954703833,
+ "grad_norm": 0.06050721928477287,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 1.9018139839172363,
+ "step": 2208
+ },
+ {
+ "epoch": 30.682926829268293,
+ "grad_norm": 0.05980826914310455,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 1.9104691743850708,
+ "step": 2209
+ },
+ {
+ "epoch": 30.696864111498257,
+ "grad_norm": 0.06286003440618515,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 1.899999737739563,
+ "step": 2210
+ },
+ {
+ "epoch": 30.71080139372822,
+ "grad_norm": 0.06179053336381912,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 1.8993868827819824,
+ "step": 2211
+ },
+ {
+ "epoch": 30.72473867595819,
+ "grad_norm": 0.06006763502955437,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 1.9067524671554565,
+ "step": 2212
+ },
+ {
+ "epoch": 30.738675958188153,
+ "grad_norm": 0.06104302033782005,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 1.9074416160583496,
+ "step": 2213
+ },
+ {
+ "epoch": 30.752613240418118,
+ "grad_norm": 0.059952687472105026,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 1.9049818515777588,
+ "step": 2214
+ },
+ {
+ "epoch": 30.766550522648085,
+ "grad_norm": 0.06469111144542694,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 1.9107921123504639,
+ "step": 2215
+ },
+ {
+ "epoch": 30.78048780487805,
+ "grad_norm": 0.06027381867170334,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 1.920630693435669,
+ "step": 2216
+ },
+ {
+ "epoch": 30.794425087108014,
+ "grad_norm": 0.06066177412867546,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 1.9131195545196533,
+ "step": 2217
+ },
+ {
+ "epoch": 30.808362369337978,
+ "grad_norm": 0.06533747911453247,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 1.9172463417053223,
+ "step": 2218
+ },
+ {
+ "epoch": 30.822299651567945,
+ "grad_norm": 0.061867427080869675,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 1.9100360870361328,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83623693379791,
+ "grad_norm": 0.06394216418266296,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 1.9218776226043701,
+ "step": 2220
+ },
+ {
+ "epoch": 30.850174216027874,
+ "grad_norm": 0.061471469700336456,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 1.901368498802185,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86411149825784,
+ "grad_norm": 0.06380587071180344,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 1.9187582731246948,
+ "step": 2222
+ },
+ {
+ "epoch": 30.878048780487806,
+ "grad_norm": 0.06244281679391861,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 1.9082577228546143,
+ "step": 2223
+ },
+ {
+ "epoch": 30.89198606271777,
+ "grad_norm": 0.06435921788215637,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 1.924039363861084,
+ "step": 2224
+ },
+ {
+ "epoch": 30.905923344947734,
+ "grad_norm": 0.06331372261047363,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 1.9083380699157715,
+ "step": 2225
+ },
+ {
+ "epoch": 30.9198606271777,
+ "grad_norm": 0.06393173336982727,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 1.9132580757141113,
+ "step": 2226
+ },
+ {
+ "epoch": 30.933797909407666,
+ "grad_norm": 0.06500784307718277,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 1.9109617471694946,
+ "step": 2227
+ },
+ {
+ "epoch": 30.94773519163763,
+ "grad_norm": 0.06389841437339783,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 1.926948070526123,
+ "step": 2228
+ },
+ {
+ "epoch": 30.961672473867594,
+ "grad_norm": 0.06343282759189606,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 1.9304101467132568,
+ "step": 2229
+ },
+ {
+ "epoch": 30.975609756097562,
+ "grad_norm": 0.06003900617361069,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 1.9066693782806396,
+ "step": 2230
+ },
+ {
+ "epoch": 30.989547038327526,
+ "grad_norm": 0.06355264037847519,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 1.9183707237243652,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0492861270904541,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 1.4322078227996826,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6370227932929993,
+ "eval_runtime": 40.5546,
+ "eval_samples_per_second": 60.215,
+ "eval_steps_per_second": 0.493,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013937282229964,
+ "grad_norm": 0.057478148490190506,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 1.887700080871582,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027874564459932,
+ "grad_norm": 0.06220356002449989,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 1.8875237703323364,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041811846689896,
+ "grad_norm": 0.05686607584357262,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 1.897495150566101,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05574912891986,
+ "grad_norm": 0.06147320941090584,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 1.8913912773132324,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069686411149824,
+ "grad_norm": 0.06075664609670639,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 1.8823286294937134,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083623693379792,
+ "grad_norm": 0.058347299695014954,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 1.8890151977539062,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097560975609756,
+ "grad_norm": 0.06388939172029495,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 1.8857264518737793,
+ "step": 2239
+ },
+ {
+ "epoch": 31.11149825783972,
+ "grad_norm": 0.05901556834578514,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 1.8779339790344238,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125435540069688,
+ "grad_norm": 0.06094517558813095,
+ "learning_rate": 0.000201141724176723,
+ "loss": 1.8811697959899902,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139372822299652,
+ "grad_norm": 0.060666318982839584,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 1.8911728858947754,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153310104529616,
+ "grad_norm": 0.05608760938048363,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 1.8804845809936523,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16724738675958,
+ "grad_norm": 0.06151990965008736,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 1.8901598453521729,
+ "step": 2244
+ },
+ {
+ "epoch": 31.181184668989548,
+ "grad_norm": 0.05805948004126549,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 1.8834660053253174,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195121951219512,
+ "grad_norm": 0.05967888608574867,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 1.8807945251464844,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209059233449477,
+ "grad_norm": 0.06295159459114075,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 1.8946118354797363,
+ "step": 2247
+ },
+ {
+ "epoch": 31.222996515679444,
+ "grad_norm": 0.057971853762865067,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 1.889994502067566,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23693379790941,
+ "grad_norm": 0.05941799655556679,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 1.884669542312622,
+ "step": 2249
+ },
+ {
+ "epoch": 31.250871080139373,
+ "grad_norm": 0.06173485890030861,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 1.8727672100067139,
+ "step": 2250
+ },
+ {
+ "epoch": 31.264808362369337,
+ "grad_norm": 0.05898866057395935,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 1.8795123100280762,
+ "step": 2251
+ },
+ {
+ "epoch": 31.278745644599304,
+ "grad_norm": 0.06072458624839783,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 1.8846805095672607,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29268292682927,
+ "grad_norm": 0.060174115002155304,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 1.8993717432022095,
+ "step": 2253
+ },
+ {
+ "epoch": 31.306620209059233,
+ "grad_norm": 0.061605289578437805,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 1.8877589702606201,
+ "step": 2254
+ },
+ {
+ "epoch": 31.320557491289197,
+ "grad_norm": 0.06246590241789818,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 1.8974242210388184,
+ "step": 2255
+ },
+ {
+ "epoch": 31.334494773519165,
+ "grad_norm": 0.060071494430303574,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 1.881445288658142,
+ "step": 2256
+ },
+ {
+ "epoch": 31.34843205574913,
+ "grad_norm": 0.060360442847013474,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 1.8938015699386597,
+ "step": 2257
+ },
+ {
+ "epoch": 31.362369337979093,
+ "grad_norm": 0.06192562356591225,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 1.8961126804351807,
+ "step": 2258
+ },
+ {
+ "epoch": 31.37630662020906,
+ "grad_norm": 0.05872832611203194,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 1.8974393606185913,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390243902439025,
+ "grad_norm": 0.05980943515896797,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 1.8886656761169434,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40418118466899,
+ "grad_norm": 0.0593055821955204,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 1.8888055086135864,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418118466898953,
+ "grad_norm": 0.059456177055835724,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 1.9045109748840332,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43205574912892,
+ "grad_norm": 0.06260113418102264,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 1.9062271118164062,
+ "step": 2263
+ },
+ {
+ "epoch": 31.445993031358885,
+ "grad_norm": 0.05866454914212227,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 1.8920798301696777,
+ "step": 2264
+ },
+ {
+ "epoch": 31.45993031358885,
+ "grad_norm": 0.060281749814748764,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 1.9048649072647095,
+ "step": 2265
+ },
+ {
+ "epoch": 31.473867595818817,
+ "grad_norm": 0.060433730483055115,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 1.8905510902404785,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48780487804878,
+ "grad_norm": 0.06074376404285431,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 1.890744686126709,
+ "step": 2267
+ },
+ {
+ "epoch": 31.501742160278745,
+ "grad_norm": 0.060131702572107315,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 1.893791913986206,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51567944250871,
+ "grad_norm": 0.05804680287837982,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 1.8838505744934082,
+ "step": 2269
+ },
+ {
+ "epoch": 31.529616724738677,
+ "grad_norm": 0.05933694168925285,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 1.897843360900879,
+ "step": 2270
+ },
+ {
+ "epoch": 31.54355400696864,
+ "grad_norm": 0.058537356555461884,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 1.8826205730438232,
+ "step": 2271
+ },
+ {
+ "epoch": 31.557491289198605,
+ "grad_norm": 0.060081496834754944,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 1.889431357383728,
+ "step": 2272
+ },
+ {
+ "epoch": 31.571428571428573,
+ "grad_norm": 0.05904357507824898,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 1.8965020179748535,
+ "step": 2273
+ },
+ {
+ "epoch": 31.585365853658537,
+ "grad_norm": 0.05924354866147041,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 1.894204020500183,
+ "step": 2274
+ },
+ {
+ "epoch": 31.5993031358885,
+ "grad_norm": 0.059307269752025604,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 1.890209436416626,
+ "step": 2275
+ },
+ {
+ "epoch": 31.613240418118465,
+ "grad_norm": 0.05969221144914627,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 1.8865056037902832,
+ "step": 2276
+ },
+ {
+ "epoch": 31.627177700348433,
+ "grad_norm": 0.060738109052181244,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 1.8825939893722534,
+ "step": 2277
+ },
+ {
+ "epoch": 31.641114982578397,
+ "grad_norm": 0.06001656875014305,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 1.8988475799560547,
+ "step": 2278
+ },
+ {
+ "epoch": 31.65505226480836,
+ "grad_norm": 0.061478059738874435,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 1.9071214199066162,
+ "step": 2279
+ },
+ {
+ "epoch": 31.66898954703833,
+ "grad_norm": 0.06089174002408981,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 1.8998515605926514,
+ "step": 2280
+ },
+ {
+ "epoch": 31.682926829268293,
+ "grad_norm": 0.05940974876284599,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 1.8907361030578613,
+ "step": 2281
+ },
+ {
+ "epoch": 31.696864111498257,
+ "grad_norm": 0.06015758961439133,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 1.9010521173477173,
+ "step": 2282
+ },
+ {
+ "epoch": 31.71080139372822,
+ "grad_norm": 0.06062757968902588,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 1.899493932723999,
+ "step": 2283
+ },
+ {
+ "epoch": 31.72473867595819,
+ "grad_norm": 0.060489553958177567,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 1.9095027446746826,
+ "step": 2284
+ },
+ {
+ "epoch": 31.738675958188153,
+ "grad_norm": 0.05949694290757179,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 1.900972604751587,
+ "step": 2285
+ },
+ {
+ "epoch": 31.752613240418118,
+ "grad_norm": 0.06269118189811707,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 1.891621708869934,
+ "step": 2286
+ },
+ {
+ "epoch": 31.766550522648085,
+ "grad_norm": 0.06033318489789963,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 1.8954113721847534,
+ "step": 2287
+ },
+ {
+ "epoch": 31.78048780487805,
+ "grad_norm": 0.0604838989675045,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 1.9047741889953613,
+ "step": 2288
+ },
+ {
+ "epoch": 31.794425087108014,
+ "grad_norm": 0.060087885707616806,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 1.8912427425384521,
+ "step": 2289
+ },
+ {
+ "epoch": 31.808362369337978,
+ "grad_norm": 0.059592097997665405,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 1.8874431848526,
+ "step": 2290
+ },
+ {
+ "epoch": 31.822299651567945,
+ "grad_norm": 0.06219378113746643,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 1.8843477964401245,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83623693379791,
+ "grad_norm": 0.06010756641626358,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 1.892686367034912,
+ "step": 2292
+ },
+ {
+ "epoch": 31.850174216027874,
+ "grad_norm": 0.061404358595609665,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 1.907541275024414,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86411149825784,
+ "grad_norm": 0.060503825545310974,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 1.891597032546997,
+ "step": 2294
+ },
+ {
+ "epoch": 31.878048780487806,
+ "grad_norm": 0.06197115406394005,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 1.8968595266342163,
+ "step": 2295
+ },
+ {
+ "epoch": 31.89198606271777,
+ "grad_norm": 0.06244310364127159,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 1.903473138809204,
+ "step": 2296
+ },
+ {
+ "epoch": 31.905923344947734,
+ "grad_norm": 0.060648202896118164,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 1.9044740200042725,
+ "step": 2297
+ },
+ {
+ "epoch": 31.9198606271777,
+ "grad_norm": 0.061299171298742294,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 1.8890066146850586,
+ "step": 2298
+ },
+ {
+ "epoch": 31.933797909407666,
+ "grad_norm": 0.06204016134142876,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 1.912150502204895,
+ "step": 2299
+ },
+ {
+ "epoch": 31.94773519163763,
+ "grad_norm": 0.06292745471000671,
+ "learning_rate": 0.000186516746349841,
+ "loss": 1.9133987426757812,
+ "step": 2300
+ },
+ {
+ "epoch": 31.961672473867594,
+ "grad_norm": 0.0633154958486557,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 1.9106671810150146,
+ "step": 2301
+ },
+ {
+ "epoch": 31.975609756097562,
+ "grad_norm": 0.06269625574350357,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 1.9090297222137451,
+ "step": 2302
+ },
+ {
+ "epoch": 31.989547038327526,
+ "grad_norm": 0.06060625612735748,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 1.8874908685684204,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.05050232633948326,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 1.4178190231323242,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6386013627052307,
+ "eval_runtime": 40.9718,
+ "eval_samples_per_second": 59.602,
+ "eval_steps_per_second": 0.488,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01393728222997,
+ "grad_norm": 0.05701690912246704,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 1.8772780895233154,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02787456445993,
+ "grad_norm": 0.05597634240984917,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 1.8772567510604858,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041811846689896,
+ "grad_norm": 0.05631265789270401,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 1.8660085201263428,
+ "step": 2307
+ },
+ {
+ "epoch": 32.055749128919864,
+ "grad_norm": 0.05604572966694832,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 1.876168966293335,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069686411149824,
+ "grad_norm": 0.056123036891222,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 1.8643467426300049,
+ "step": 2309
+ },
+ {
+ "epoch": 32.08362369337979,
+ "grad_norm": 0.05791968107223511,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 1.866549015045166,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09756097560975,
+ "grad_norm": 0.05521981045603752,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 1.8852781057357788,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11149825783972,
+ "grad_norm": 0.055954866111278534,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 1.8603641986846924,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12543554006969,
+ "grad_norm": 0.05931234359741211,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 1.8619375228881836,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13937282229965,
+ "grad_norm": 0.05743015930056572,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 1.8824023008346558,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153310104529616,
+ "grad_norm": 0.056393202394247055,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 1.8722079992294312,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167247386759584,
+ "grad_norm": 0.05671187862753868,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 1.8824248313903809,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181184668989545,
+ "grad_norm": 0.05667170509696007,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 1.882739782333374,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19512195121951,
+ "grad_norm": 0.05515167862176895,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 1.8555771112442017,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20905923344948,
+ "grad_norm": 0.05649147555232048,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 1.8673419952392578,
+ "step": 2319
+ },
+ {
+ "epoch": 32.22299651567944,
+ "grad_norm": 0.058623913675546646,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 1.8744664192199707,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23693379790941,
+ "grad_norm": 0.055131133645772934,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 1.8781135082244873,
+ "step": 2321
+ },
+ {
+ "epoch": 32.250871080139376,
+ "grad_norm": 0.06008831411600113,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 1.8844444751739502,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26480836236934,
+ "grad_norm": 0.05755964294075966,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 1.8790829181671143,
+ "step": 2323
+ },
+ {
+ "epoch": 32.278745644599304,
+ "grad_norm": 0.055803194642066956,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 1.8642692565917969,
+ "step": 2324
+ },
+ {
+ "epoch": 32.292682926829265,
+ "grad_norm": 0.05710228905081749,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 1.8659579753875732,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30662020905923,
+ "grad_norm": 0.05771065130829811,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 1.8812355995178223,
+ "step": 2326
+ },
+ {
+ "epoch": 32.3205574912892,
+ "grad_norm": 0.05642757564783096,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 1.8610783815383911,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33449477351916,
+ "grad_norm": 0.05687689036130905,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 1.875876545906067,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34843205574913,
+ "grad_norm": 0.057835984975099564,
+ "learning_rate": 0.000179445406945268,
+ "loss": 1.8674073219299316,
+ "step": 2329
+ },
+ {
+ "epoch": 32.362369337979096,
+ "grad_norm": 0.05800129100680351,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 1.8708086013793945,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37630662020906,
+ "grad_norm": 0.05744084715843201,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 1.8869245052337646,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390243902439025,
+ "grad_norm": 0.056703388690948486,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 1.8688050508499146,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40418118466899,
+ "grad_norm": 0.06032602861523628,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 1.873382568359375,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41811846689895,
+ "grad_norm": 0.05810776352882385,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 1.8624725341796875,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43205574912892,
+ "grad_norm": 0.060990724712610245,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 1.8792011737823486,
+ "step": 2335
+ },
+ {
+ "epoch": 32.44599303135889,
+ "grad_norm": 0.05975554883480072,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 1.8793275356292725,
+ "step": 2336
+ },
+ {
+ "epoch": 32.45993031358885,
+ "grad_norm": 0.05772533640265465,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 1.889489769935608,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47386759581882,
+ "grad_norm": 0.05920211225748062,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 1.8647074699401855,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48780487804878,
+ "grad_norm": 0.059745173901319504,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 1.8649309873580933,
+ "step": 2339
+ },
+ {
+ "epoch": 32.501742160278745,
+ "grad_norm": 0.05933282524347305,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 1.8936223983764648,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51567944250871,
+ "grad_norm": 0.062407393008470535,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 1.8831725120544434,
+ "step": 2341
+ },
+ {
+ "epoch": 32.52961672473867,
+ "grad_norm": 0.061185337603092194,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 1.8766106367111206,
+ "step": 2342
+ },
+ {
+ "epoch": 32.54355400696864,
+ "grad_norm": 0.06082421913743019,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 1.880276083946228,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55749128919861,
+ "grad_norm": 0.05900120362639427,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 1.8696876764297485,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57142857142857,
+ "grad_norm": 0.06146271899342537,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 1.864588737487793,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58536585365854,
+ "grad_norm": 0.05976773053407669,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 1.8771488666534424,
+ "step": 2346
+ },
+ {
+ "epoch": 32.599303135888505,
+ "grad_norm": 0.05808904021978378,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 1.890838623046875,
+ "step": 2347
+ },
+ {
+ "epoch": 32.613240418118465,
+ "grad_norm": 0.06149672716856003,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 1.8787143230438232,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62717770034843,
+ "grad_norm": 0.059028271585702896,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 1.8875718116760254,
+ "step": 2349
+ },
+ {
+ "epoch": 32.641114982578394,
+ "grad_norm": 0.062163304537534714,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 1.86748206615448,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65505226480836,
+ "grad_norm": 0.06221195310354233,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 1.8742306232452393,
+ "step": 2351
+ },
+ {
+ "epoch": 32.66898954703833,
+ "grad_norm": 0.060461003333330154,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 1.885633945465088,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68292682926829,
+ "grad_norm": 0.06130106747150421,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 1.8827314376831055,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69686411149826,
+ "grad_norm": 0.06044832617044449,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 1.8724548816680908,
+ "step": 2354
+ },
+ {
+ "epoch": 32.710801393728225,
+ "grad_norm": 0.06029713526368141,
+ "learning_rate": 0.000173176617304673,
+ "loss": 1.8745346069335938,
+ "step": 2355
+ },
+ {
+ "epoch": 32.724738675958186,
+ "grad_norm": 0.06287439167499542,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 1.8879387378692627,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73867595818815,
+ "grad_norm": 0.059636496007442474,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 1.882830262184143,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75261324041812,
+ "grad_norm": 0.06590992957353592,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 1.882242202758789,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76655052264808,
+ "grad_norm": 0.06283026933670044,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 1.8864011764526367,
+ "step": 2359
+ },
+ {
+ "epoch": 32.78048780487805,
+ "grad_norm": 0.061717696487903595,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 1.8815748691558838,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79442508710802,
+ "grad_norm": 0.06423293799161911,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 1.8824026584625244,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80836236933798,
+ "grad_norm": 0.06351540237665176,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 1.9000264406204224,
+ "step": 2362
+ },
+ {
+ "epoch": 32.822299651567945,
+ "grad_norm": 0.060434527695178986,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 1.8875732421875,
+ "step": 2363
+ },
+ {
+ "epoch": 32.836236933797906,
+ "grad_norm": 0.062305059283971786,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 1.8695244789123535,
+ "step": 2364
+ },
+ {
+ "epoch": 32.850174216027874,
+ "grad_norm": 0.05981886014342308,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 1.8885269165039062,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86411149825784,
+ "grad_norm": 0.0642906129360199,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 1.8818528652191162,
+ "step": 2366
+ },
+ {
+ "epoch": 32.8780487804878,
+ "grad_norm": 0.06316667050123215,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 1.8994319438934326,
+ "step": 2367
+ },
+ {
+ "epoch": 32.89198606271777,
+ "grad_norm": 0.058260831981897354,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 1.8908791542053223,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90592334494774,
+ "grad_norm": 0.062237389385700226,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 1.8753025531768799,
+ "step": 2369
+ },
+ {
+ "epoch": 32.9198606271777,
+ "grad_norm": 0.05972493812441826,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 1.899425745010376,
+ "step": 2370
+ },
+ {
+ "epoch": 32.933797909407666,
+ "grad_norm": 0.058846332132816315,
+ "learning_rate": 0.00016935382741164,
+ "loss": 1.8934335708618164,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94773519163763,
+ "grad_norm": 0.06333424896001816,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 1.8951395750045776,
+ "step": 2372
+ },
+ {
+ "epoch": 32.961672473867594,
+ "grad_norm": 0.06063693389296532,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 1.881688117980957,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97560975609756,
+ "grad_norm": 0.060873813927173615,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 1.8836770057678223,
+ "step": 2374
+ },
+ {
+ "epoch": 32.98954703832753,
+ "grad_norm": 0.060378748923540115,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 1.886624813079834,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.04906065762042999,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 1.4146361351013184,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6401442885398865,
+ "eval_runtime": 41.0841,
+ "eval_samples_per_second": 59.439,
+ "eval_steps_per_second": 0.487,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01393728222997,
+ "grad_norm": 0.05557359382510185,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 1.8605964183807373,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02787456445993,
+ "grad_norm": 0.055585991591215134,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 1.8456363677978516,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041811846689896,
+ "grad_norm": 0.05610295757651329,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 1.847888469696045,
+ "step": 2379
+ },
+ {
+ "epoch": 33.055749128919864,
+ "grad_norm": 0.05532662197947502,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 1.850597620010376,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069686411149824,
+ "grad_norm": 0.05585319921374321,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 1.8694689273834229,
+ "step": 2381
+ },
+ {
+ "epoch": 33.08362369337979,
+ "grad_norm": 0.05658521130681038,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 1.8533908128738403,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09756097560975,
+ "grad_norm": 0.0558413490653038,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 1.8735582828521729,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11149825783972,
+ "grad_norm": 0.05637028068304062,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 1.8585509061813354,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12543554006969,
+ "grad_norm": 0.055854618549346924,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 1.8539409637451172,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13937282229965,
+ "grad_norm": 0.055990077555179596,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 1.85886812210083,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153310104529616,
+ "grad_norm": 0.05781752988696098,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 1.8603525161743164,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167247386759584,
+ "grad_norm": 0.055855557322502136,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 1.861933946609497,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181184668989545,
+ "grad_norm": 0.05729927122592926,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 1.86094069480896,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19512195121951,
+ "grad_norm": 0.056495532393455505,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 1.848013997077942,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20905923344948,
+ "grad_norm": 0.0571073479950428,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 1.8558173179626465,
+ "step": 2391
+ },
+ {
+ "epoch": 33.22299651567944,
+ "grad_norm": 0.056066226214170456,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 1.8723382949829102,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23693379790941,
+ "grad_norm": 0.05570382624864578,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 1.864349365234375,
+ "step": 2393
+ },
+ {
+ "epoch": 33.250871080139376,
+ "grad_norm": 0.054912492632865906,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 1.8678479194641113,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26480836236934,
+ "grad_norm": 0.05622183158993721,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 1.8648667335510254,
+ "step": 2395
+ },
+ {
+ "epoch": 33.278745644599304,
+ "grad_norm": 0.05719740688800812,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 1.850623607635498,
+ "step": 2396
+ },
+ {
+ "epoch": 33.292682926829265,
+ "grad_norm": 0.056594349443912506,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 1.8707754611968994,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30662020905923,
+ "grad_norm": 0.05596613511443138,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 1.8580527305603027,
+ "step": 2398
+ },
+ {
+ "epoch": 33.3205574912892,
+ "grad_norm": 0.05675627663731575,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 1.8687434196472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33449477351916,
+ "grad_norm": 0.05572087690234184,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 1.8603391647338867,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34843205574913,
+ "grad_norm": 0.05837463587522507,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 1.8638341426849365,
+ "step": 2401
+ },
+ {
+ "epoch": 33.362369337979096,
+ "grad_norm": 0.056136030703783035,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 1.853350281715393,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37630662020906,
+ "grad_norm": 0.0555158406496048,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 1.8775548934936523,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390243902439025,
+ "grad_norm": 0.0595950111746788,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 1.8578253984451294,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40418118466899,
+ "grad_norm": 0.05565760284662247,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 1.8508214950561523,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41811846689895,
+ "grad_norm": 0.058832235634326935,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 1.8614635467529297,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43205574912892,
+ "grad_norm": 0.0585586316883564,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 1.8617370128631592,
+ "step": 2407
+ },
+ {
+ "epoch": 33.44599303135889,
+ "grad_norm": 0.05951899662613869,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 1.8648651838302612,
+ "step": 2408
+ },
+ {
+ "epoch": 33.45993031358885,
+ "grad_norm": 0.057204652577638626,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 1.8577271699905396,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47386759581882,
+ "grad_norm": 0.05844349041581154,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 1.8560724258422852,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48780487804878,
+ "grad_norm": 0.05761651322245598,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 1.863309383392334,
+ "step": 2411
+ },
+ {
+ "epoch": 33.501742160278745,
+ "grad_norm": 0.05818328261375427,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 1.8798553943634033,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51567944250871,
+ "grad_norm": 0.05746658146381378,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 1.8628616333007812,
+ "step": 2413
+ },
+ {
+ "epoch": 33.52961672473867,
+ "grad_norm": 0.057656314224004745,
+ "learning_rate": 0.000159218843594243,
+ "loss": 1.8699405193328857,
+ "step": 2414
+ },
+ {
+ "epoch": 33.54355400696864,
+ "grad_norm": 0.0586664155125618,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 1.8481391668319702,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55749128919861,
+ "grad_norm": 0.05769997090101242,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 1.8612866401672363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57142857142857,
+ "grad_norm": 0.0598456896841526,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 1.8580410480499268,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58536585365854,
+ "grad_norm": 0.05761442333459854,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 1.874891996383667,
+ "step": 2418
+ },
+ {
+ "epoch": 33.599303135888505,
+ "grad_norm": 0.06176742538809776,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 1.8660908937454224,
+ "step": 2419
+ },
+ {
+ "epoch": 33.613240418118465,
+ "grad_norm": 0.0588943213224411,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 1.8718886375427246,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62717770034843,
+ "grad_norm": 0.06092666834592819,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 1.8634107112884521,
+ "step": 2421
+ },
+ {
+ "epoch": 33.641114982578394,
+ "grad_norm": 0.05774684250354767,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 1.868361234664917,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65505226480836,
+ "grad_norm": 0.05932781100273132,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 1.867889642715454,
+ "step": 2423
+ },
+ {
+ "epoch": 33.66898954703833,
+ "grad_norm": 0.058603718876838684,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 1.8793346881866455,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68292682926829,
+ "grad_norm": 0.05870316922664642,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 1.8717231750488281,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69686411149826,
+ "grad_norm": 0.05805273354053497,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 1.8479297161102295,
+ "step": 2426
+ },
+ {
+ "epoch": 33.710801393728225,
+ "grad_norm": 0.059018444269895554,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 1.8711276054382324,
+ "step": 2427
+ },
+ {
+ "epoch": 33.724738675958186,
+ "grad_norm": 0.05927985534071922,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 1.8653428554534912,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73867595818815,
+ "grad_norm": 0.06085766851902008,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 1.8724851608276367,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75261324041812,
+ "grad_norm": 0.05974132940173149,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 1.8459011316299438,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76655052264808,
+ "grad_norm": 0.05851755291223526,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 1.863109827041626,
+ "step": 2431
+ },
+ {
+ "epoch": 33.78048780487805,
+ "grad_norm": 0.06079421192407608,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 1.8687033653259277,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79442508710802,
+ "grad_norm": 0.057930078357458115,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 1.8653396368026733,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80836236933798,
+ "grad_norm": 0.059440065175294876,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 1.8569486141204834,
+ "step": 2434
+ },
+ {
+ "epoch": 33.822299651567945,
+ "grad_norm": 0.05835104361176491,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 1.8828535079956055,
+ "step": 2435
+ },
+ {
+ "epoch": 33.836236933797906,
+ "grad_norm": 0.06108623370528221,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 1.8749947547912598,
+ "step": 2436
+ },
+ {
+ "epoch": 33.850174216027874,
+ "grad_norm": 0.059921979904174805,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 1.8684660196304321,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86411149825784,
+ "grad_norm": 0.059964295476675034,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 1.862084150314331,
+ "step": 2438
+ },
+ {
+ "epoch": 33.8780487804878,
+ "grad_norm": 0.05888533219695091,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 1.85898756980896,
+ "step": 2439
+ },
+ {
+ "epoch": 33.89198606271777,
+ "grad_norm": 0.058946020901203156,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 1.872513771057129,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90592334494774,
+ "grad_norm": 0.05947734788060188,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 1.8610495328903198,
+ "step": 2441
+ },
+ {
+ "epoch": 33.9198606271777,
+ "grad_norm": 0.05859588086605072,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 1.8842005729675293,
+ "step": 2442
+ },
+ {
+ "epoch": 33.933797909407666,
+ "grad_norm": 0.05818016827106476,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 1.8715083599090576,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94773519163763,
+ "grad_norm": 0.05960588902235031,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 1.881840467453003,
+ "step": 2444
+ },
+ {
+ "epoch": 33.961672473867594,
+ "grad_norm": 0.059536516666412354,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 1.8734736442565918,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97560975609756,
+ "grad_norm": 0.05922378972172737,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 1.8569964170455933,
+ "step": 2446
+ },
+ {
+ "epoch": 33.98954703832753,
+ "grad_norm": 0.058573752641677856,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 1.8710730075836182,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.047325484454631805,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 1.4022152423858643,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6414204835891724,
+ "eval_runtime": 39.9489,
+ "eval_samples_per_second": 61.128,
+ "eval_steps_per_second": 0.501,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01393728222997,
+ "grad_norm": 0.05358244478702545,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 1.8384610414505005,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02787456445993,
+ "grad_norm": 0.0558239221572876,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 1.8538508415222168,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041811846689896,
+ "grad_norm": 0.05390622094273567,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 1.8460195064544678,
+ "step": 2451
+ },
+ {
+ "epoch": 34.055749128919864,
+ "grad_norm": 0.056224383413791656,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 1.8416059017181396,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069686411149824,
+ "grad_norm": 0.053925771266222,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 1.8532803058624268,
+ "step": 2453
+ },
+ {
+ "epoch": 34.08362369337979,
+ "grad_norm": 0.05750057101249695,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 1.847919225692749,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09756097560975,
+ "grad_norm": 0.05499042198061943,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 1.8450297117233276,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11149825783972,
+ "grad_norm": 0.0544770173728466,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 1.8482413291931152,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12543554006969,
+ "grad_norm": 0.05656695365905762,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 1.860766053199768,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13937282229965,
+ "grad_norm": 0.054617058485746384,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 1.8398511409759521,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153310104529616,
+ "grad_norm": 0.05470099300146103,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 1.8501665592193604,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167247386759584,
+ "grad_norm": 0.05437156930565834,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 1.8514611721038818,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181184668989545,
+ "grad_norm": 0.05481892451643944,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 1.8424533605575562,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19512195121951,
+ "grad_norm": 0.055016402155160904,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 1.823396921157837,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20905923344948,
+ "grad_norm": 0.05608086287975311,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 1.8525217771530151,
+ "step": 2463
+ },
+ {
+ "epoch": 34.22299651567944,
+ "grad_norm": 0.055000413209199905,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 1.8570690155029297,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23693379790941,
+ "grad_norm": 0.05437581241130829,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 1.8471980094909668,
+ "step": 2465
+ },
+ {
+ "epoch": 34.250871080139376,
+ "grad_norm": 0.0553814135491848,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 1.8384971618652344,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26480836236934,
+ "grad_norm": 0.053538259118795395,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 1.8536994457244873,
+ "step": 2467
+ },
+ {
+ "epoch": 34.278745644599304,
+ "grad_norm": 0.05509314313530922,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 1.8386962413787842,
+ "step": 2468
+ },
+ {
+ "epoch": 34.292682926829265,
+ "grad_norm": 0.054552506655454636,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 1.8502833843231201,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30662020905923,
+ "grad_norm": 0.054326511919498444,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 1.8350508213043213,
+ "step": 2470
+ },
+ {
+ "epoch": 34.3205574912892,
+ "grad_norm": 0.05458610877394676,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 1.8488143682479858,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33449477351916,
+ "grad_norm": 0.05537666752934456,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 1.8561567068099976,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34843205574913,
+ "grad_norm": 0.056333623826503754,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 1.8610734939575195,
+ "step": 2473
+ },
+ {
+ "epoch": 34.362369337979096,
+ "grad_norm": 0.053708501160144806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 1.8342688083648682,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37630662020906,
+ "grad_norm": 0.05496754124760628,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 1.8592373132705688,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390243902439025,
+ "grad_norm": 0.05445944517850876,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 1.8551034927368164,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40418118466899,
+ "grad_norm": 0.055870212614536285,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 1.8510769605636597,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41811846689895,
+ "grad_norm": 0.05536561831831932,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 1.8463027477264404,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43205574912892,
+ "grad_norm": 0.054308436810970306,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 1.8470419645309448,
+ "step": 2479
+ },
+ {
+ "epoch": 34.44599303135889,
+ "grad_norm": 0.056683722883462906,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 1.8605597019195557,
+ "step": 2480
+ },
+ {
+ "epoch": 34.45993031358885,
+ "grad_norm": 0.05601316690444946,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 1.847717046737671,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47386759581882,
+ "grad_norm": 0.05449012294411659,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 1.8464031219482422,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48780487804878,
+ "grad_norm": 0.054641272872686386,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 1.832083821296692,
+ "step": 2483
+ },
+ {
+ "epoch": 34.501742160278745,
+ "grad_norm": 0.05610528588294983,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 1.8700275421142578,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51567944250871,
+ "grad_norm": 0.0551300011575222,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 1.8514257669448853,
+ "step": 2485
+ },
+ {
+ "epoch": 34.52961672473867,
+ "grad_norm": 0.056593794375658035,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 1.8480379581451416,
+ "step": 2486
+ },
+ {
+ "epoch": 34.54355400696864,
+ "grad_norm": 0.055673424154520035,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 1.838927984237671,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55749128919861,
+ "grad_norm": 0.05493558943271637,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 1.8383840322494507,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57142857142857,
+ "grad_norm": 0.05680295452475548,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 1.8514118194580078,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58536585365854,
+ "grad_norm": 0.057038743048906326,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 1.8620257377624512,
+ "step": 2490
+ },
+ {
+ "epoch": 34.599303135888505,
+ "grad_norm": 0.05735227093100548,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 1.8413968086242676,
+ "step": 2491
+ },
+ {
+ "epoch": 34.613240418118465,
+ "grad_norm": 0.058245740830898285,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 1.8599750995635986,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62717770034843,
+ "grad_norm": 0.056244585663080215,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 1.8527486324310303,
+ "step": 2493
+ },
+ {
+ "epoch": 34.641114982578394,
+ "grad_norm": 0.05616258084774017,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 1.84233558177948,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65505226480836,
+ "grad_norm": 0.057303108274936676,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 1.865498661994934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.66898954703833,
+ "grad_norm": 0.05895329639315605,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 1.8521645069122314,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68292682926829,
+ "grad_norm": 0.05679212138056755,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 1.8539199829101562,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69686411149826,
+ "grad_norm": 0.05897561088204384,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 1.8534815311431885,
+ "step": 2498
+ },
+ {
+ "epoch": 34.710801393728225,
+ "grad_norm": 0.05891786143183708,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 1.845531702041626,
+ "step": 2499
+ },
+ {
+ "epoch": 34.724738675958186,
+ "grad_norm": 0.059177134186029434,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 1.8502740859985352,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73867595818815,
+ "grad_norm": 0.05862346291542053,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 1.8657329082489014,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75261324041812,
+ "grad_norm": 0.061109986156225204,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 1.8452383279800415,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76655052264808,
+ "grad_norm": 0.058431629091501236,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 1.8575074672698975,
+ "step": 2503
+ },
+ {
+ "epoch": 34.78048780487805,
+ "grad_norm": 0.05970519408583641,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 1.8476191759109497,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79442508710802,
+ "grad_norm": 0.05823173746466637,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 1.8554785251617432,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80836236933798,
+ "grad_norm": 0.060084082186222076,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 1.8471784591674805,
+ "step": 2506
+ },
+ {
+ "epoch": 34.822299651567945,
+ "grad_norm": 0.0589512400329113,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 1.86192786693573,
+ "step": 2507
+ },
+ {
+ "epoch": 34.836236933797906,
+ "grad_norm": 0.06184513866901398,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 1.8572009801864624,
+ "step": 2508
+ },
+ {
+ "epoch": 34.850174216027874,
+ "grad_norm": 0.05833953991532326,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 1.8561553955078125,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86411149825784,
+ "grad_norm": 0.061172984540462494,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 1.8583234548568726,
+ "step": 2510
+ },
+ {
+ "epoch": 34.8780487804878,
+ "grad_norm": 0.0573224313557148,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 1.8403494358062744,
+ "step": 2511
+ },
+ {
+ "epoch": 34.89198606271777,
+ "grad_norm": 0.06135294586420059,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 1.869114875793457,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90592334494774,
+ "grad_norm": 0.05907885357737541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 1.8581886291503906,
+ "step": 2513
+ },
+ {
+ "epoch": 34.9198606271777,
+ "grad_norm": 0.06035003438591957,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 1.8454012870788574,
+ "step": 2514
+ },
+ {
+ "epoch": 34.933797909407666,
+ "grad_norm": 0.060585640370845795,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 1.8453309535980225,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94773519163763,
+ "grad_norm": 0.06050759181380272,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 1.8467144966125488,
+ "step": 2516
+ },
+ {
+ "epoch": 34.961672473867594,
+ "grad_norm": 0.06118239462375641,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 1.8623638153076172,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97560975609756,
+ "grad_norm": 0.059003446251153946,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 1.8540363311767578,
+ "step": 2518
+ },
+ {
+ "epoch": 34.98954703832753,
+ "grad_norm": 0.06078164279460907,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 1.8434364795684814,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04768386855721474,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 1.3819507360458374,
+ "step": 2520
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0771796057495962e+18,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2520/training_args.bin b/runs/mt5-ssl/checkpoint-2520/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2520/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-2880/chat_template.jinja b/runs/mt5-ssl/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-2880/config.json b/runs/mt5-ssl/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-2880/generation_config.json b/runs/mt5-ssl/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-2880/model.safetensors b/runs/mt5-ssl/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f3c73a108d82e957efc91609c254f5f798e06907
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:298e80bf1749453cddd11a87428827efd3a822fb192c23476ae3ce40bb267229
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-2880/optimizer.pt b/runs/mt5-ssl/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5e3ca5e30e85bb431c250efa308c34a85053f3a5
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:876db827bb2be2bfe34e5560cc8046a6afc0b35ac9db321918f17c121bb1c3f3
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-2880/rng_state_0.pth b/runs/mt5-ssl/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b73f9a380e4e9c61520ae0674c9453be2b007f8d
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:19e0f78f2d9435829a3e127bfda2ca14f3245c5620079efefde656536bbe279e
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2880/rng_state_1.pth b/runs/mt5-ssl/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1c80ff79c801a37a1192dfdb25ad4404ef9b73aa
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0ff567f23d2e2493cacddea05b5da8eee7e39fe5f45b9a2a98aa6a743f71addc
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-2880/scheduler.pt b/runs/mt5-ssl/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f1b2d6bb8ce3cef13aa84e428e5db83abeeff7b5
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6c45747ad72424e6ab5eef3447cbd13933161ee71d5ad9aaac15ca345390629c
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-2880/tokenizer.json b/runs/mt5-ssl/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-2880/tokenizer_config.json b/runs/mt5-ssl/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2880/trainer_state.json b/runs/mt5-ssl/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..b128067ad5f7696c2a6f962d76763b7dca29cef0
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20506 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6287577152252197,
+ "eval_runtime": 39.8508,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013937282229964,
+ "grad_norm": 0.056210409849882126,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 1.9758458137512207,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027874564459932,
+ "grad_norm": 0.06387963145971298,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 1.9743283987045288,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041811846689896,
+ "grad_norm": 0.059931688010692596,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 1.983224630355835,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05574912891986,
+ "grad_norm": 0.059698957949876785,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 1.9638850688934326,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069686411149824,
+ "grad_norm": 0.057556964457035065,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 1.9888463020324707,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083623693379792,
+ "grad_norm": 0.06250642985105515,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 1.9776880741119385,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097560975609756,
+ "grad_norm": 0.06218443810939789,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 1.9904823303222656,
+ "step": 1807
+ },
+ {
+ "epoch": 25.11149825783972,
+ "grad_norm": 0.06147411838173866,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 1.9881136417388916,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125435540069688,
+ "grad_norm": 0.06080890819430351,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 1.9833264350891113,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139372822299652,
+ "grad_norm": 0.057465583086013794,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 1.9764835834503174,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153310104529616,
+ "grad_norm": 0.061488356441259384,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 1.9807426929473877,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16724738675958,
+ "grad_norm": 0.060737013816833496,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 1.9796113967895508,
+ "step": 1812
+ },
+ {
+ "epoch": 25.181184668989548,
+ "grad_norm": 0.060508858412504196,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 1.9862518310546875,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195121951219512,
+ "grad_norm": 0.06062212213873863,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 1.9855456352233887,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209059233449477,
+ "grad_norm": 0.06143258139491081,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 1.9763805866241455,
+ "step": 1815
+ },
+ {
+ "epoch": 25.222996515679444,
+ "grad_norm": 0.06446489691734314,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 1.9674923419952393,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23693379790941,
+ "grad_norm": 0.06179432198405266,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 1.9792906045913696,
+ "step": 1817
+ },
+ {
+ "epoch": 25.250871080139373,
+ "grad_norm": 0.062219955027103424,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 1.983303189277649,
+ "step": 1818
+ },
+ {
+ "epoch": 25.264808362369337,
+ "grad_norm": 0.060631707310676575,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 1.9859293699264526,
+ "step": 1819
+ },
+ {
+ "epoch": 25.278745644599304,
+ "grad_norm": 0.06329842656850815,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 1.988373041152954,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29268292682927,
+ "grad_norm": 0.06290270388126373,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 1.9817759990692139,
+ "step": 1821
+ },
+ {
+ "epoch": 25.306620209059233,
+ "grad_norm": 0.062234487384557724,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 1.9615445137023926,
+ "step": 1822
+ },
+ {
+ "epoch": 25.320557491289197,
+ "grad_norm": 0.06063779070973396,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 1.9736756086349487,
+ "step": 1823
+ },
+ {
+ "epoch": 25.334494773519165,
+ "grad_norm": 0.06245938315987587,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 1.9932905435562134,
+ "step": 1824
+ },
+ {
+ "epoch": 25.34843205574913,
+ "grad_norm": 0.06380622833967209,
+ "learning_rate": 0.000309930621834911,
+ "loss": 1.9695440530776978,
+ "step": 1825
+ },
+ {
+ "epoch": 25.362369337979093,
+ "grad_norm": 0.0626329705119133,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 1.9820970296859741,
+ "step": 1826
+ },
+ {
+ "epoch": 25.37630662020906,
+ "grad_norm": 0.060844022780656815,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 1.9926050901412964,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390243902439025,
+ "grad_norm": 0.06239933893084526,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 2.001865863800049,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40418118466899,
+ "grad_norm": 0.06491564959287643,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 1.9897377490997314,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418118466898953,
+ "grad_norm": 0.05946580320596695,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 1.973170280456543,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43205574912892,
+ "grad_norm": 0.05930647253990173,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 1.9949123859405518,
+ "step": 1831
+ },
+ {
+ "epoch": 25.445993031358885,
+ "grad_norm": 0.06269685924053192,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 1.9912991523742676,
+ "step": 1832
+ },
+ {
+ "epoch": 25.45993031358885,
+ "grad_norm": 0.0637395903468132,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 1.9869414567947388,
+ "step": 1833
+ },
+ {
+ "epoch": 25.473867595818817,
+ "grad_norm": 0.0628778263926506,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 1.9748289585113525,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48780487804878,
+ "grad_norm": 0.06064455583691597,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 2.008216381072998,
+ "step": 1835
+ },
+ {
+ "epoch": 25.501742160278745,
+ "grad_norm": 0.0655415952205658,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 1.9852811098098755,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51567944250871,
+ "grad_norm": 0.06189798563718796,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 1.983716368675232,
+ "step": 1837
+ },
+ {
+ "epoch": 25.529616724738677,
+ "grad_norm": 0.061454080045223236,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 2.013745069503784,
+ "step": 1838
+ },
+ {
+ "epoch": 25.54355400696864,
+ "grad_norm": 0.06712181121110916,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 1.9856178760528564,
+ "step": 1839
+ },
+ {
+ "epoch": 25.557491289198605,
+ "grad_norm": 0.06364461034536362,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 2.0034379959106445,
+ "step": 1840
+ },
+ {
+ "epoch": 25.571428571428573,
+ "grad_norm": 0.06669855862855911,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 1.9868637323379517,
+ "step": 1841
+ },
+ {
+ "epoch": 25.585365853658537,
+ "grad_norm": 0.06578591465950012,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 1.9998185634613037,
+ "step": 1842
+ },
+ {
+ "epoch": 25.5993031358885,
+ "grad_norm": 0.06305381655693054,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 2.000594139099121,
+ "step": 1843
+ },
+ {
+ "epoch": 25.613240418118465,
+ "grad_norm": 0.061369556933641434,
+ "learning_rate": 0.000304866093757771,
+ "loss": 1.9811584949493408,
+ "step": 1844
+ },
+ {
+ "epoch": 25.627177700348433,
+ "grad_norm": 0.062390975654125214,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 2.0020859241485596,
+ "step": 1845
+ },
+ {
+ "epoch": 25.641114982578397,
+ "grad_norm": 0.06477145850658417,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 2.004000186920166,
+ "step": 1846
+ },
+ {
+ "epoch": 25.65505226480836,
+ "grad_norm": 0.06288585811853409,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 1.9919252395629883,
+ "step": 1847
+ },
+ {
+ "epoch": 25.66898954703833,
+ "grad_norm": 0.05959830805659294,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 1.9860963821411133,
+ "step": 1848
+ },
+ {
+ "epoch": 25.682926829268293,
+ "grad_norm": 0.0596182718873024,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 2.0014610290527344,
+ "step": 1849
+ },
+ {
+ "epoch": 25.696864111498257,
+ "grad_norm": 0.06665527820587158,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 2.0072989463806152,
+ "step": 1850
+ },
+ {
+ "epoch": 25.71080139372822,
+ "grad_norm": 0.0631684735417366,
+ "learning_rate": 0.000303,
+ "loss": 2.0076632499694824,
+ "step": 1851
+ },
+ {
+ "epoch": 25.72473867595819,
+ "grad_norm": 0.06214272975921631,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 2.0143537521362305,
+ "step": 1852
+ },
+ {
+ "epoch": 25.738675958188153,
+ "grad_norm": 0.06422685086727142,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 1.9885808229446411,
+ "step": 1853
+ },
+ {
+ "epoch": 25.752613240418118,
+ "grad_norm": 0.06583613157272339,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 2.0042953491210938,
+ "step": 1854
+ },
+ {
+ "epoch": 25.766550522648085,
+ "grad_norm": 0.06265029311180115,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 2.003157138824463,
+ "step": 1855
+ },
+ {
+ "epoch": 25.78048780487805,
+ "grad_norm": 0.0612902007997036,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 1.9832231998443604,
+ "step": 1856
+ },
+ {
+ "epoch": 25.794425087108014,
+ "grad_norm": 0.0625016912817955,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 2.0012407302856445,
+ "step": 1857
+ },
+ {
+ "epoch": 25.808362369337978,
+ "grad_norm": 0.062357913702726364,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 2.010338306427002,
+ "step": 1858
+ },
+ {
+ "epoch": 25.822299651567945,
+ "grad_norm": 0.06349784880876541,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 2.0105648040771484,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83623693379791,
+ "grad_norm": 0.061822179704904556,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 1.9987019300460815,
+ "step": 1860
+ },
+ {
+ "epoch": 25.850174216027874,
+ "grad_norm": 0.06443580985069275,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 1.994943618774414,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86411149825784,
+ "grad_norm": 0.061674103140830994,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 2.0116522312164307,
+ "step": 1862
+ },
+ {
+ "epoch": 25.878048780487806,
+ "grad_norm": 0.061944279819726944,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 2.012296438217163,
+ "step": 1863
+ },
+ {
+ "epoch": 25.89198606271777,
+ "grad_norm": 0.06270956248044968,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 2.0057339668273926,
+ "step": 1864
+ },
+ {
+ "epoch": 25.905923344947734,
+ "grad_norm": 0.06427973508834839,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 2.0053863525390625,
+ "step": 1865
+ },
+ {
+ "epoch": 25.9198606271777,
+ "grad_norm": 0.0637369304895401,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 2.0130667686462402,
+ "step": 1866
+ },
+ {
+ "epoch": 25.933797909407666,
+ "grad_norm": 0.06291835755109787,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 2.0094404220581055,
+ "step": 1867
+ },
+ {
+ "epoch": 25.94773519163763,
+ "grad_norm": 0.06374962627887726,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 2.0074503421783447,
+ "step": 1868
+ },
+ {
+ "epoch": 25.961672473867594,
+ "grad_norm": 0.06330475956201553,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 1.9999325275421143,
+ "step": 1869
+ },
+ {
+ "epoch": 25.975609756097562,
+ "grad_norm": 0.06077918782830238,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 1.994345784187317,
+ "step": 1870
+ },
+ {
+ "epoch": 25.989547038327526,
+ "grad_norm": 0.06524437665939331,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 2.0210394859313965,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.05064176768064499,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 1.4825201034545898,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6298588514328003,
+ "eval_runtime": 43.1768,
+ "eval_samples_per_second": 56.558,
+ "eval_steps_per_second": 0.463,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013937282229964,
+ "grad_norm": 0.05861504375934601,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 1.9728282690048218,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027874564459932,
+ "grad_norm": 0.06309985369443893,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 1.9638245105743408,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041811846689896,
+ "grad_norm": 0.06438111513853073,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 1.9697308540344238,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05574912891986,
+ "grad_norm": 0.060332708060741425,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 1.9574062824249268,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069686411149824,
+ "grad_norm": 0.061463311314582825,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 1.9469952583312988,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083623693379792,
+ "grad_norm": 0.06337480992078781,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 1.9695435762405396,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097560975609756,
+ "grad_norm": 0.06303273886442184,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 1.9764964580535889,
+ "step": 1879
+ },
+ {
+ "epoch": 26.11149825783972,
+ "grad_norm": 0.06061594933271408,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 1.9758808612823486,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125435540069688,
+ "grad_norm": 0.061181873083114624,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 1.9607892036437988,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139372822299652,
+ "grad_norm": 0.06557024270296097,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 1.9752333164215088,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153310104529616,
+ "grad_norm": 0.06261707097291946,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 1.9634147882461548,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16724738675958,
+ "grad_norm": 0.05993996560573578,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 1.957287311553955,
+ "step": 1884
+ },
+ {
+ "epoch": 26.181184668989548,
+ "grad_norm": 0.0612766332924366,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 1.9562265872955322,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195121951219512,
+ "grad_norm": 0.05735815688967705,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 1.9577300548553467,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209059233449477,
+ "grad_norm": 0.06258611381053925,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 1.9661660194396973,
+ "step": 1887
+ },
+ {
+ "epoch": 26.222996515679444,
+ "grad_norm": 0.06298062205314636,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 1.9508070945739746,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23693379790941,
+ "grad_norm": 0.06346822530031204,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 1.9677343368530273,
+ "step": 1889
+ },
+ {
+ "epoch": 26.250871080139373,
+ "grad_norm": 0.05959425866603851,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 1.9574894905090332,
+ "step": 1890
+ },
+ {
+ "epoch": 26.264808362369337,
+ "grad_norm": 0.0610729418694973,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 1.9515372514724731,
+ "step": 1891
+ },
+ {
+ "epoch": 26.278745644599304,
+ "grad_norm": 0.06362602859735489,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 1.966064453125,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29268292682927,
+ "grad_norm": 0.06124011054635048,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 1.9654541015625,
+ "step": 1893
+ },
+ {
+ "epoch": 26.306620209059233,
+ "grad_norm": 0.05773317068815231,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 1.9610934257507324,
+ "step": 1894
+ },
+ {
+ "epoch": 26.320557491289197,
+ "grad_norm": 0.05896785855293274,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 1.965133547782898,
+ "step": 1895
+ },
+ {
+ "epoch": 26.334494773519165,
+ "grad_norm": 0.059504859149456024,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 1.9617180824279785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.34843205574913,
+ "grad_norm": 0.05804497376084328,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 1.9761323928833008,
+ "step": 1897
+ },
+ {
+ "epoch": 26.362369337979093,
+ "grad_norm": 0.058211296796798706,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 1.9805936813354492,
+ "step": 1898
+ },
+ {
+ "epoch": 26.37630662020906,
+ "grad_norm": 0.05976752191781998,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 1.9514508247375488,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390243902439025,
+ "grad_norm": 0.05824121832847595,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 1.97285795211792,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40418118466899,
+ "grad_norm": 0.05793534219264984,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 1.9608209133148193,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418118466898953,
+ "grad_norm": 0.05697625130414963,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 1.9787089824676514,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43205574912892,
+ "grad_norm": 0.05703677609562874,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 1.9610021114349365,
+ "step": 1903
+ },
+ {
+ "epoch": 26.445993031358885,
+ "grad_norm": 0.057380311191082,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 1.9764081239700317,
+ "step": 1904
+ },
+ {
+ "epoch": 26.45993031358885,
+ "grad_norm": 0.05846763774752617,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 1.9641292095184326,
+ "step": 1905
+ },
+ {
+ "epoch": 26.473867595818817,
+ "grad_norm": 0.05929148569703102,
+ "learning_rate": 0.000288343693342466,
+ "loss": 1.9803684949874878,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48780487804878,
+ "grad_norm": 0.058236971497535706,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 1.981757402420044,
+ "step": 1907
+ },
+ {
+ "epoch": 26.501742160278745,
+ "grad_norm": 0.05845103785395622,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 1.9736125469207764,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51567944250871,
+ "grad_norm": 0.06104041263461113,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 1.9698188304901123,
+ "step": 1909
+ },
+ {
+ "epoch": 26.529616724738677,
+ "grad_norm": 0.0628688707947731,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 1.9786568880081177,
+ "step": 1910
+ },
+ {
+ "epoch": 26.54355400696864,
+ "grad_norm": 0.06115978956222534,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 1.9744799137115479,
+ "step": 1911
+ },
+ {
+ "epoch": 26.557491289198605,
+ "grad_norm": 0.06233362853527069,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 1.9801995754241943,
+ "step": 1912
+ },
+ {
+ "epoch": 26.571428571428573,
+ "grad_norm": 0.06182500347495079,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 1.9888432025909424,
+ "step": 1913
+ },
+ {
+ "epoch": 26.585365853658537,
+ "grad_norm": 0.05962767079472542,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 1.9648957252502441,
+ "step": 1914
+ },
+ {
+ "epoch": 26.5993031358885,
+ "grad_norm": 0.05994252488017082,
+ "learning_rate": 0.000285947841605349,
+ "loss": 1.9709041118621826,
+ "step": 1915
+ },
+ {
+ "epoch": 26.613240418118465,
+ "grad_norm": 0.06045769527554512,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 1.994588017463684,
+ "step": 1916
+ },
+ {
+ "epoch": 26.627177700348433,
+ "grad_norm": 0.059465229511260986,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 1.9743266105651855,
+ "step": 1917
+ },
+ {
+ "epoch": 26.641114982578397,
+ "grad_norm": 0.06020190939307213,
+ "learning_rate": 0.000285149463934261,
+ "loss": 1.9780995845794678,
+ "step": 1918
+ },
+ {
+ "epoch": 26.65505226480836,
+ "grad_norm": 0.05829813703894615,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 1.9831392765045166,
+ "step": 1919
+ },
+ {
+ "epoch": 26.66898954703833,
+ "grad_norm": 0.05960174649953842,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 1.9799871444702148,
+ "step": 1920
+ },
+ {
+ "epoch": 26.682926829268293,
+ "grad_norm": 0.06133116036653519,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 1.985158920288086,
+ "step": 1921
+ },
+ {
+ "epoch": 26.696864111498257,
+ "grad_norm": 0.05947015434503555,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 1.9783923625946045,
+ "step": 1922
+ },
+ {
+ "epoch": 26.71080139372822,
+ "grad_norm": 0.061687882989645004,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 1.9685261249542236,
+ "step": 1923
+ },
+ {
+ "epoch": 26.72473867595819,
+ "grad_norm": 0.0613567978143692,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 1.9945828914642334,
+ "step": 1924
+ },
+ {
+ "epoch": 26.738675958188153,
+ "grad_norm": 0.05902605876326561,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 1.966120719909668,
+ "step": 1925
+ },
+ {
+ "epoch": 26.752613240418118,
+ "grad_norm": 0.06018664687871933,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 1.991615653038025,
+ "step": 1926
+ },
+ {
+ "epoch": 26.766550522648085,
+ "grad_norm": 0.061171986162662506,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 1.9877536296844482,
+ "step": 1927
+ },
+ {
+ "epoch": 26.78048780487805,
+ "grad_norm": 0.06263386458158493,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 1.9799100160598755,
+ "step": 1928
+ },
+ {
+ "epoch": 26.794425087108014,
+ "grad_norm": 0.06326548755168915,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 1.9805757999420166,
+ "step": 1929
+ },
+ {
+ "epoch": 26.808362369337978,
+ "grad_norm": 0.06226228550076485,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 1.9867990016937256,
+ "step": 1930
+ },
+ {
+ "epoch": 26.822299651567945,
+ "grad_norm": 0.06003205478191376,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 1.9632174968719482,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83623693379791,
+ "grad_norm": 0.06075739488005638,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 1.9638378620147705,
+ "step": 1932
+ },
+ {
+ "epoch": 26.850174216027874,
+ "grad_norm": 0.061726443469524384,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 1.9865589141845703,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86411149825784,
+ "grad_norm": 0.0621049702167511,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 2.007071018218994,
+ "step": 1934
+ },
+ {
+ "epoch": 26.878048780487806,
+ "grad_norm": 0.061793744564056396,
+ "learning_rate": 0.000280627938758204,
+ "loss": 1.988924264907837,
+ "step": 1935
+ },
+ {
+ "epoch": 26.89198606271777,
+ "grad_norm": 0.062355563044548035,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 1.9830653667449951,
+ "step": 1936
+ },
+ {
+ "epoch": 26.905923344947734,
+ "grad_norm": 0.060881514102220535,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 2.003551959991455,
+ "step": 1937
+ },
+ {
+ "epoch": 26.9198606271777,
+ "grad_norm": 0.061703141778707504,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 1.9811534881591797,
+ "step": 1938
+ },
+ {
+ "epoch": 26.933797909407666,
+ "grad_norm": 0.0634910985827446,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 1.9840846061706543,
+ "step": 1939
+ },
+ {
+ "epoch": 26.94773519163763,
+ "grad_norm": 0.060807306319475174,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 1.9879541397094727,
+ "step": 1940
+ },
+ {
+ "epoch": 26.961672473867594,
+ "grad_norm": 0.06126011535525322,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 1.993988037109375,
+ "step": 1941
+ },
+ {
+ "epoch": 26.975609756097562,
+ "grad_norm": 0.06394185870885849,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 1.983365535736084,
+ "step": 1942
+ },
+ {
+ "epoch": 26.989547038327526,
+ "grad_norm": 0.06164516508579254,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 1.989990234375,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.04997039958834648,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 1.4921388626098633,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6310494542121887,
+ "eval_runtime": 40.875,
+ "eval_samples_per_second": 59.743,
+ "eval_steps_per_second": 0.489,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013937282229964,
+ "grad_norm": 0.05894223228096962,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 1.935828447341919,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027874564459932,
+ "grad_norm": 0.06731119751930237,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 1.945769190788269,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041811846689896,
+ "grad_norm": 0.06787799298763275,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 1.944061040878296,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05574912891986,
+ "grad_norm": 0.05840576812624931,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 1.94246506690979,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069686411149824,
+ "grad_norm": 0.058151841163635254,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 1.940029263496399,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083623693379792,
+ "grad_norm": 0.06329379975795746,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 1.9523427486419678,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097560975609756,
+ "grad_norm": 0.06530784070491791,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 1.9352362155914307,
+ "step": 1951
+ },
+ {
+ "epoch": 27.11149825783972,
+ "grad_norm": 0.06273649632930756,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 1.946252703666687,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125435540069688,
+ "grad_norm": 0.05996445193886757,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 1.9437150955200195,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139372822299652,
+ "grad_norm": 0.059087123721838,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 1.936135172843933,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153310104529616,
+ "grad_norm": 0.06257616728544235,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 1.9452544450759888,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16724738675958,
+ "grad_norm": 0.06591545045375824,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 1.9334893226623535,
+ "step": 1956
+ },
+ {
+ "epoch": 27.181184668989548,
+ "grad_norm": 0.0595872700214386,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 1.9514997005462646,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195121951219512,
+ "grad_norm": 0.05855429172515869,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 1.955148458480835,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209059233449477,
+ "grad_norm": 0.06383495032787323,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 1.9607688188552856,
+ "step": 1959
+ },
+ {
+ "epoch": 27.222996515679444,
+ "grad_norm": 0.06486682593822479,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 1.941009283065796,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23693379790941,
+ "grad_norm": 0.06293730437755585,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 1.9550879001617432,
+ "step": 1961
+ },
+ {
+ "epoch": 27.250871080139373,
+ "grad_norm": 0.06104227155447006,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 1.9483157396316528,
+ "step": 1962
+ },
+ {
+ "epoch": 27.264808362369337,
+ "grad_norm": 0.06920639425516129,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 1.948688268661499,
+ "step": 1963
+ },
+ {
+ "epoch": 27.278745644599304,
+ "grad_norm": 0.06186635419726372,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 1.9512240886688232,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29268292682927,
+ "grad_norm": 0.06258274614810944,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 1.9538028240203857,
+ "step": 1965
+ },
+ {
+ "epoch": 27.306620209059233,
+ "grad_norm": 0.06546445935964584,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 1.9497663974761963,
+ "step": 1966
+ },
+ {
+ "epoch": 27.320557491289197,
+ "grad_norm": 0.06289096921682358,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 1.9706182479858398,
+ "step": 1967
+ },
+ {
+ "epoch": 27.334494773519165,
+ "grad_norm": 0.06412214040756226,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 1.9423331022262573,
+ "step": 1968
+ },
+ {
+ "epoch": 27.34843205574913,
+ "grad_norm": 0.062773697078228,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 1.9471882581710815,
+ "step": 1969
+ },
+ {
+ "epoch": 27.362369337979093,
+ "grad_norm": 0.060290850698947906,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 1.9732742309570312,
+ "step": 1970
+ },
+ {
+ "epoch": 27.37630662020906,
+ "grad_norm": 0.062357380986213684,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 1.957550048828125,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390243902439025,
+ "grad_norm": 0.06070727854967117,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 1.951438069343567,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40418118466899,
+ "grad_norm": 0.06380726397037506,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 1.9512609243392944,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418118466898953,
+ "grad_norm": 0.06135795637965202,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 1.9663273096084595,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43205574912892,
+ "grad_norm": 0.0627567246556282,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 1.964203119277954,
+ "step": 1975
+ },
+ {
+ "epoch": 27.445993031358885,
+ "grad_norm": 0.06235141307115555,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 1.960247278213501,
+ "step": 1976
+ },
+ {
+ "epoch": 27.45993031358885,
+ "grad_norm": 0.06187676638364792,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 1.9667054414749146,
+ "step": 1977
+ },
+ {
+ "epoch": 27.473867595818817,
+ "grad_norm": 0.06090894341468811,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 1.9695777893066406,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48780487804878,
+ "grad_norm": 0.06363680213689804,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 1.9529361724853516,
+ "step": 1979
+ },
+ {
+ "epoch": 27.501742160278745,
+ "grad_norm": 0.06053898110985756,
+ "learning_rate": 0.00026868712586269,
+ "loss": 1.9707324504852295,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51567944250871,
+ "grad_norm": 0.061363156884908676,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 1.955542802810669,
+ "step": 1981
+ },
+ {
+ "epoch": 27.529616724738677,
+ "grad_norm": 0.06158224865794182,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 1.9489967823028564,
+ "step": 1982
+ },
+ {
+ "epoch": 27.54355400696864,
+ "grad_norm": 0.05886901170015335,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 1.9705257415771484,
+ "step": 1983
+ },
+ {
+ "epoch": 27.557491289198605,
+ "grad_norm": 0.05890485271811485,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 1.9430618286132812,
+ "step": 1984
+ },
+ {
+ "epoch": 27.571428571428573,
+ "grad_norm": 0.060433268547058105,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 1.9634191989898682,
+ "step": 1985
+ },
+ {
+ "epoch": 27.585365853658537,
+ "grad_norm": 0.0626354068517685,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 1.9568289518356323,
+ "step": 1986
+ },
+ {
+ "epoch": 27.5993031358885,
+ "grad_norm": 0.05973317474126816,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 1.9678292274475098,
+ "step": 1987
+ },
+ {
+ "epoch": 27.613240418118465,
+ "grad_norm": 0.062289632856845856,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 1.9413363933563232,
+ "step": 1988
+ },
+ {
+ "epoch": 27.627177700348433,
+ "grad_norm": 0.06162180379033089,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 1.9835751056671143,
+ "step": 1989
+ },
+ {
+ "epoch": 27.641114982578397,
+ "grad_norm": 0.06176188960671425,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 1.9541220664978027,
+ "step": 1990
+ },
+ {
+ "epoch": 27.65505226480836,
+ "grad_norm": 0.06479815393686295,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 1.9596972465515137,
+ "step": 1991
+ },
+ {
+ "epoch": 27.66898954703833,
+ "grad_norm": 0.06571854650974274,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 1.9628980159759521,
+ "step": 1992
+ },
+ {
+ "epoch": 27.682926829268293,
+ "grad_norm": 0.06167568638920784,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 1.9624769687652588,
+ "step": 1993
+ },
+ {
+ "epoch": 27.696864111498257,
+ "grad_norm": 0.06207694113254547,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 1.9487217664718628,
+ "step": 1994
+ },
+ {
+ "epoch": 27.71080139372822,
+ "grad_norm": 0.061430424451828,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 1.9567784070968628,
+ "step": 1995
+ },
+ {
+ "epoch": 27.72473867595819,
+ "grad_norm": 0.06589192897081375,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 1.9594354629516602,
+ "step": 1996
+ },
+ {
+ "epoch": 27.738675958188153,
+ "grad_norm": 0.06640033423900604,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 1.9637835025787354,
+ "step": 1997
+ },
+ {
+ "epoch": 27.752613240418118,
+ "grad_norm": 0.06181980296969414,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 1.9597129821777344,
+ "step": 1998
+ },
+ {
+ "epoch": 27.766550522648085,
+ "grad_norm": 0.06352445483207703,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 1.9612562656402588,
+ "step": 1999
+ },
+ {
+ "epoch": 27.78048780487805,
+ "grad_norm": 0.06398574262857437,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 1.9662930965423584,
+ "step": 2000
+ },
+ {
+ "epoch": 27.794425087108014,
+ "grad_norm": 0.06195084750652313,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 1.956035852432251,
+ "step": 2001
+ },
+ {
+ "epoch": 27.808362369337978,
+ "grad_norm": 0.05951351299881935,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 1.9712438583374023,
+ "step": 2002
+ },
+ {
+ "epoch": 27.822299651567945,
+ "grad_norm": 0.062185633927583694,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 1.9714539051055908,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83623693379791,
+ "grad_norm": 0.061178892850875854,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 1.9691438674926758,
+ "step": 2004
+ },
+ {
+ "epoch": 27.850174216027874,
+ "grad_norm": 0.06262431293725967,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 1.9506231546401978,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86411149825784,
+ "grad_norm": 0.06206081807613373,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 1.9662809371948242,
+ "step": 2006
+ },
+ {
+ "epoch": 27.878048780487806,
+ "grad_norm": 0.0626990869641304,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 1.972696304321289,
+ "step": 2007
+ },
+ {
+ "epoch": 27.89198606271777,
+ "grad_norm": 0.06345485150814056,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 1.9560344219207764,
+ "step": 2008
+ },
+ {
+ "epoch": 27.905923344947734,
+ "grad_norm": 0.06401040405035019,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 1.9732205867767334,
+ "step": 2009
+ },
+ {
+ "epoch": 27.9198606271777,
+ "grad_norm": 0.06437943130731583,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 1.974897027015686,
+ "step": 2010
+ },
+ {
+ "epoch": 27.933797909407666,
+ "grad_norm": 0.06296383589506149,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 1.9710102081298828,
+ "step": 2011
+ },
+ {
+ "epoch": 27.94773519163763,
+ "grad_norm": 0.06527025997638702,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 1.9707660675048828,
+ "step": 2012
+ },
+ {
+ "epoch": 27.961672473867594,
+ "grad_norm": 0.06185369938611984,
+ "learning_rate": 0.00025996500713765,
+ "loss": 1.9675114154815674,
+ "step": 2013
+ },
+ {
+ "epoch": 27.975609756097562,
+ "grad_norm": 0.06357269734144211,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 1.9812211990356445,
+ "step": 2014
+ },
+ {
+ "epoch": 27.989547038327526,
+ "grad_norm": 0.06581240892410278,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 1.97153639793396,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.049184102565050125,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 1.4647765159606934,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6328383088111877,
+ "eval_runtime": 40.5927,
+ "eval_samples_per_second": 60.159,
+ "eval_steps_per_second": 0.493,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013937282229964,
+ "grad_norm": 0.05999046564102173,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 1.9182333946228027,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027874564459932,
+ "grad_norm": 0.059804853051900864,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 1.924762487411499,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041811846689896,
+ "grad_norm": 0.05890650674700737,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 1.9341628551483154,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05574912891986,
+ "grad_norm": 0.057233694940805435,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 1.932476282119751,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069686411149824,
+ "grad_norm": 0.060101717710494995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 1.9403120279312134,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083623693379792,
+ "grad_norm": 0.06054912135004997,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 1.9274934530258179,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097560975609756,
+ "grad_norm": 0.05882031098008156,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 1.9331915378570557,
+ "step": 2023
+ },
+ {
+ "epoch": 28.11149825783972,
+ "grad_norm": 0.05846688151359558,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 1.9342529773712158,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125435540069688,
+ "grad_norm": 0.058974213898181915,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 1.9353152513504028,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139372822299652,
+ "grad_norm": 0.05882132798433304,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 1.9500072002410889,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153310104529616,
+ "grad_norm": 0.05919627472758293,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 1.9313960075378418,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16724738675958,
+ "grad_norm": 0.06413958966732025,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 1.9516873359680176,
+ "step": 2028
+ },
+ {
+ "epoch": 28.181184668989548,
+ "grad_norm": 0.05799189209938049,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 1.9290452003479004,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195121951219512,
+ "grad_norm": 0.05999152734875679,
+ "learning_rate": 0.000255486047794226,
+ "loss": 1.9396628141403198,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209059233449477,
+ "grad_norm": 0.060240138322114944,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 1.937988519668579,
+ "step": 2031
+ },
+ {
+ "epoch": 28.222996515679444,
+ "grad_norm": 0.05933740735054016,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 1.9343246221542358,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23693379790941,
+ "grad_norm": 0.05960093438625336,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 1.9326903820037842,
+ "step": 2033
+ },
+ {
+ "epoch": 28.250871080139373,
+ "grad_norm": 0.0635334923863411,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 1.9381543397903442,
+ "step": 2034
+ },
+ {
+ "epoch": 28.264808362369337,
+ "grad_norm": 0.06162877380847931,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 1.9265705347061157,
+ "step": 2035
+ },
+ {
+ "epoch": 28.278745644599304,
+ "grad_norm": 0.059690047055482864,
+ "learning_rate": 0.000253907826333243,
+ "loss": 1.9278099536895752,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29268292682927,
+ "grad_norm": 0.0612013153731823,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 1.939026117324829,
+ "step": 2037
+ },
+ {
+ "epoch": 28.306620209059233,
+ "grad_norm": 0.06080419942736626,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 1.9306461811065674,
+ "step": 2038
+ },
+ {
+ "epoch": 28.320557491289197,
+ "grad_norm": 0.06308622658252716,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 1.9350788593292236,
+ "step": 2039
+ },
+ {
+ "epoch": 28.334494773519165,
+ "grad_norm": 0.06124545633792877,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 1.9264733791351318,
+ "step": 2040
+ },
+ {
+ "epoch": 28.34843205574913,
+ "grad_norm": 0.059037938714027405,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 1.917725920677185,
+ "step": 2041
+ },
+ {
+ "epoch": 28.362369337979093,
+ "grad_norm": 0.06089222431182861,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 1.949327826499939,
+ "step": 2042
+ },
+ {
+ "epoch": 28.37630662020906,
+ "grad_norm": 0.05957084521651268,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 1.938978910446167,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390243902439025,
+ "grad_norm": 0.06060048192739487,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 1.9226362705230713,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40418118466899,
+ "grad_norm": 0.058304425328969955,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 1.9330644607543945,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418118466898953,
+ "grad_norm": 0.05970550701022148,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 1.931627631187439,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43205574912892,
+ "grad_norm": 0.05994785949587822,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 1.9387202262878418,
+ "step": 2047
+ },
+ {
+ "epoch": 28.445993031358885,
+ "grad_norm": 0.06001518666744232,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 1.951300859451294,
+ "step": 2048
+ },
+ {
+ "epoch": 28.45993031358885,
+ "grad_norm": 0.0636107549071312,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 1.9358865022659302,
+ "step": 2049
+ },
+ {
+ "epoch": 28.473867595818817,
+ "grad_norm": 0.059402115643024445,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 1.9298653602600098,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48780487804878,
+ "grad_norm": 0.06260179728269577,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 1.954633116722107,
+ "step": 2051
+ },
+ {
+ "epoch": 28.501742160278745,
+ "grad_norm": 0.06439191848039627,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 1.9338306188583374,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51567944250871,
+ "grad_norm": 0.06265310943126678,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 1.9500043392181396,
+ "step": 2053
+ },
+ {
+ "epoch": 28.529616724738677,
+ "grad_norm": 0.06323659420013428,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 1.9445586204528809,
+ "step": 2054
+ },
+ {
+ "epoch": 28.54355400696864,
+ "grad_norm": 0.06512445956468582,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 1.958587408065796,
+ "step": 2055
+ },
+ {
+ "epoch": 28.557491289198605,
+ "grad_norm": 0.06386571377515793,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 1.949887990951538,
+ "step": 2056
+ },
+ {
+ "epoch": 28.571428571428573,
+ "grad_norm": 0.06317705661058426,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 1.9512674808502197,
+ "step": 2057
+ },
+ {
+ "epoch": 28.585365853658537,
+ "grad_norm": 0.06504984200000763,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 1.9242960214614868,
+ "step": 2058
+ },
+ {
+ "epoch": 28.5993031358885,
+ "grad_norm": 0.06278593838214874,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 1.9203131198883057,
+ "step": 2059
+ },
+ {
+ "epoch": 28.613240418118465,
+ "grad_norm": 0.06171257048845291,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 1.954156517982483,
+ "step": 2060
+ },
+ {
+ "epoch": 28.627177700348433,
+ "grad_norm": 0.06303241103887558,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 1.9460458755493164,
+ "step": 2061
+ },
+ {
+ "epoch": 28.641114982578397,
+ "grad_norm": 0.06252635270357132,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 1.9477293491363525,
+ "step": 2062
+ },
+ {
+ "epoch": 28.65505226480836,
+ "grad_norm": 0.06226412579417229,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 1.9259647130966187,
+ "step": 2063
+ },
+ {
+ "epoch": 28.66898954703833,
+ "grad_norm": 0.06111626327037811,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 1.9573581218719482,
+ "step": 2064
+ },
+ {
+ "epoch": 28.682926829268293,
+ "grad_norm": 0.06460461020469666,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 1.9486393928527832,
+ "step": 2065
+ },
+ {
+ "epoch": 28.696864111498257,
+ "grad_norm": 0.062027834355831146,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 1.9404425621032715,
+ "step": 2066
+ },
+ {
+ "epoch": 28.71080139372822,
+ "grad_norm": 0.061942074447870255,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 1.9444060325622559,
+ "step": 2067
+ },
+ {
+ "epoch": 28.72473867595819,
+ "grad_norm": 0.06159934028983116,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 1.9635517597198486,
+ "step": 2068
+ },
+ {
+ "epoch": 28.738675958188153,
+ "grad_norm": 0.0621839202940464,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 1.9442765712738037,
+ "step": 2069
+ },
+ {
+ "epoch": 28.752613240418118,
+ "grad_norm": 0.06432677060365677,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 1.9604840278625488,
+ "step": 2070
+ },
+ {
+ "epoch": 28.766550522648085,
+ "grad_norm": 0.06037316843867302,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 1.948632836341858,
+ "step": 2071
+ },
+ {
+ "epoch": 28.78048780487805,
+ "grad_norm": 0.05999939888715744,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 1.9451751708984375,
+ "step": 2072
+ },
+ {
+ "epoch": 28.794425087108014,
+ "grad_norm": 0.06441698968410492,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 1.9356517791748047,
+ "step": 2073
+ },
+ {
+ "epoch": 28.808362369337978,
+ "grad_norm": 0.06059470772743225,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 1.9258331060409546,
+ "step": 2074
+ },
+ {
+ "epoch": 28.822299651567945,
+ "grad_norm": 0.062317874282598495,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 1.9342000484466553,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83623693379791,
+ "grad_norm": 0.06318258494138718,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 1.9395548105239868,
+ "step": 2076
+ },
+ {
+ "epoch": 28.850174216027874,
+ "grad_norm": 0.06204817444086075,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 1.9572374820709229,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86411149825784,
+ "grad_norm": 0.06250188499689102,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 1.9465739727020264,
+ "step": 2078
+ },
+ {
+ "epoch": 28.878048780487806,
+ "grad_norm": 0.06293781846761703,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 1.9431169033050537,
+ "step": 2079
+ },
+ {
+ "epoch": 28.89198606271777,
+ "grad_norm": 0.059349458664655685,
+ "learning_rate": 0.000242380656502223,
+ "loss": 1.9673197269439697,
+ "step": 2080
+ },
+ {
+ "epoch": 28.905923344947734,
+ "grad_norm": 0.06214660778641701,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 1.9646787643432617,
+ "step": 2081
+ },
+ {
+ "epoch": 28.9198606271777,
+ "grad_norm": 0.062279123812913895,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 1.9572803974151611,
+ "step": 2082
+ },
+ {
+ "epoch": 28.933797909407666,
+ "grad_norm": 0.06502588093280792,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 1.9510462284088135,
+ "step": 2083
+ },
+ {
+ "epoch": 28.94773519163763,
+ "grad_norm": 0.06153101101517677,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 1.9507923126220703,
+ "step": 2084
+ },
+ {
+ "epoch": 28.961672473867594,
+ "grad_norm": 0.06469229608774185,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 1.937751293182373,
+ "step": 2085
+ },
+ {
+ "epoch": 28.975609756097562,
+ "grad_norm": 0.0672568678855896,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 1.944124698638916,
+ "step": 2086
+ },
+ {
+ "epoch": 28.989547038327526,
+ "grad_norm": 0.06451335549354553,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 1.9519702196121216,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.05141456052660942,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 1.4445865154266357,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6343450546264648,
+ "eval_runtime": 40.5081,
+ "eval_samples_per_second": 60.284,
+ "eval_steps_per_second": 0.494,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013937282229964,
+ "grad_norm": 0.061562713235616684,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 1.921946406364441,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027874564459932,
+ "grad_norm": 0.06396839767694473,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 1.9222416877746582,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041811846689896,
+ "grad_norm": 0.05828692391514778,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 1.9173905849456787,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05574912891986,
+ "grad_norm": 0.06567556411027908,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 1.9148579835891724,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069686411149824,
+ "grad_norm": 0.05924173817038536,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 1.916333794593811,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083623693379792,
+ "grad_norm": 0.06129243224859238,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 1.9031038284301758,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097560975609756,
+ "grad_norm": 0.06068028509616852,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 1.9165983200073242,
+ "step": 2095
+ },
+ {
+ "epoch": 29.11149825783972,
+ "grad_norm": 0.06304074823856354,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 1.90432870388031,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125435540069688,
+ "grad_norm": 0.060179345309734344,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 1.9187151193618774,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139372822299652,
+ "grad_norm": 0.05633116513490677,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 1.9125087261199951,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153310104529616,
+ "grad_norm": 0.0598532073199749,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 1.906031847000122,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16724738675958,
+ "grad_norm": 0.05987119674682617,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 1.9177607297897339,
+ "step": 2100
+ },
+ {
+ "epoch": 29.181184668989548,
+ "grad_norm": 0.060883522033691406,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 1.9008433818817139,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195121951219512,
+ "grad_norm": 0.05977628752589226,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 1.9176266193389893,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209059233449477,
+ "grad_norm": 0.05939047038555145,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 1.9227800369262695,
+ "step": 2103
+ },
+ {
+ "epoch": 29.222996515679444,
+ "grad_norm": 0.059280600398778915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 1.9164855480194092,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23693379790941,
+ "grad_norm": 0.058746229857206345,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 1.9110995531082153,
+ "step": 2105
+ },
+ {
+ "epoch": 29.250871080139373,
+ "grad_norm": 0.05882417410612106,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 1.926957368850708,
+ "step": 2106
+ },
+ {
+ "epoch": 29.264808362369337,
+ "grad_norm": 0.05827188491821289,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 1.9101874828338623,
+ "step": 2107
+ },
+ {
+ "epoch": 29.278745644599304,
+ "grad_norm": 0.059739116579294205,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 1.924944519996643,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29268292682927,
+ "grad_norm": 0.058013755828142166,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 1.9121344089508057,
+ "step": 2109
+ },
+ {
+ "epoch": 29.306620209059233,
+ "grad_norm": 0.061207350343465805,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 1.9137616157531738,
+ "step": 2110
+ },
+ {
+ "epoch": 29.320557491289197,
+ "grad_norm": 0.05848725140094757,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 1.9290939569473267,
+ "step": 2111
+ },
+ {
+ "epoch": 29.334494773519165,
+ "grad_norm": 0.059134867042303085,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 1.919262409210205,
+ "step": 2112
+ },
+ {
+ "epoch": 29.34843205574913,
+ "grad_norm": 0.06025470420718193,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 1.9021251201629639,
+ "step": 2113
+ },
+ {
+ "epoch": 29.362369337979093,
+ "grad_norm": 0.05981536582112312,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 1.9163634777069092,
+ "step": 2114
+ },
+ {
+ "epoch": 29.37630662020906,
+ "grad_norm": 0.05826934054493904,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 1.8987491130828857,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390243902439025,
+ "grad_norm": 0.06095297262072563,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 1.9322285652160645,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40418118466899,
+ "grad_norm": 0.05814153701066971,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 1.927016019821167,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418118466898953,
+ "grad_norm": 0.05872226879000664,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 1.9271752834320068,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43205574912892,
+ "grad_norm": 0.06008293107151985,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 1.9269304275512695,
+ "step": 2119
+ },
+ {
+ "epoch": 29.445993031358885,
+ "grad_norm": 0.06035081297159195,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 1.929135799407959,
+ "step": 2120
+ },
+ {
+ "epoch": 29.45993031358885,
+ "grad_norm": 0.06080561876296997,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 1.9284312725067139,
+ "step": 2121
+ },
+ {
+ "epoch": 29.473867595818817,
+ "grad_norm": 0.058793697506189346,
+ "learning_rate": 0.000231465389734324,
+ "loss": 1.931719183921814,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48780487804878,
+ "grad_norm": 0.06385456770658493,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 1.9112051725387573,
+ "step": 2123
+ },
+ {
+ "epoch": 29.501742160278745,
+ "grad_norm": 0.06030426174402237,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 1.9254870414733887,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51567944250871,
+ "grad_norm": 0.059933315962553024,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 1.915220022201538,
+ "step": 2125
+ },
+ {
+ "epoch": 29.529616724738677,
+ "grad_norm": 0.06083494424819946,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 1.9391965866088867,
+ "step": 2126
+ },
+ {
+ "epoch": 29.54355400696864,
+ "grad_norm": 0.06049460917711258,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 1.9301774501800537,
+ "step": 2127
+ },
+ {
+ "epoch": 29.557491289198605,
+ "grad_norm": 0.05953316017985344,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 1.9150071144104004,
+ "step": 2128
+ },
+ {
+ "epoch": 29.571428571428573,
+ "grad_norm": 0.0611971914768219,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 1.9265213012695312,
+ "step": 2129
+ },
+ {
+ "epoch": 29.585365853658537,
+ "grad_norm": 0.06123765930533409,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 1.9405107498168945,
+ "step": 2130
+ },
+ {
+ "epoch": 29.5993031358885,
+ "grad_norm": 0.06403640657663345,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 1.9216606616973877,
+ "step": 2131
+ },
+ {
+ "epoch": 29.613240418118465,
+ "grad_norm": 0.06530711054801941,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 1.9302153587341309,
+ "step": 2132
+ },
+ {
+ "epoch": 29.627177700348433,
+ "grad_norm": 0.06106860563158989,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 1.930979609489441,
+ "step": 2133
+ },
+ {
+ "epoch": 29.641114982578397,
+ "grad_norm": 0.06148209050297737,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 1.9276363849639893,
+ "step": 2134
+ },
+ {
+ "epoch": 29.65505226480836,
+ "grad_norm": 0.060283321887254715,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 1.9269459247589111,
+ "step": 2135
+ },
+ {
+ "epoch": 29.66898954703833,
+ "grad_norm": 0.06138703227043152,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 1.9389945268630981,
+ "step": 2136
+ },
+ {
+ "epoch": 29.682926829268293,
+ "grad_norm": 0.06101245433092117,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 1.9426372051239014,
+ "step": 2137
+ },
+ {
+ "epoch": 29.696864111498257,
+ "grad_norm": 0.05958810821175575,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 1.9327162504196167,
+ "step": 2138
+ },
+ {
+ "epoch": 29.71080139372822,
+ "grad_norm": 0.06250876933336258,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 1.9384305477142334,
+ "step": 2139
+ },
+ {
+ "epoch": 29.72473867595819,
+ "grad_norm": 0.06344381719827652,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 1.9422640800476074,
+ "step": 2140
+ },
+ {
+ "epoch": 29.738675958188153,
+ "grad_norm": 0.06065250560641289,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 1.920068621635437,
+ "step": 2141
+ },
+ {
+ "epoch": 29.752613240418118,
+ "grad_norm": 0.06333960592746735,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 1.925317645072937,
+ "step": 2142
+ },
+ {
+ "epoch": 29.766550522648085,
+ "grad_norm": 0.06285370141267776,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 1.92912757396698,
+ "step": 2143
+ },
+ {
+ "epoch": 29.78048780487805,
+ "grad_norm": 0.06441807746887207,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 1.941272497177124,
+ "step": 2144
+ },
+ {
+ "epoch": 29.794425087108014,
+ "grad_norm": 0.06314355880022049,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 1.923574447631836,
+ "step": 2145
+ },
+ {
+ "epoch": 29.808362369337978,
+ "grad_norm": 0.0639389157295227,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 1.9339193105697632,
+ "step": 2146
+ },
+ {
+ "epoch": 29.822299651567945,
+ "grad_norm": 0.06419187039136887,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 1.9294497966766357,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83623693379791,
+ "grad_norm": 0.06351246684789658,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 1.9282407760620117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.850174216027874,
+ "grad_norm": 0.06371650099754333,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 1.914009690284729,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86411149825784,
+ "grad_norm": 0.06891132146120071,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 1.948460578918457,
+ "step": 2150
+ },
+ {
+ "epoch": 29.878048780487806,
+ "grad_norm": 0.06575452536344528,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 1.9238896369934082,
+ "step": 2151
+ },
+ {
+ "epoch": 29.89198606271777,
+ "grad_norm": 0.06279128044843674,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 1.9414188861846924,
+ "step": 2152
+ },
+ {
+ "epoch": 29.905923344947734,
+ "grad_norm": 0.06493876129388809,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 1.9337334632873535,
+ "step": 2153
+ },
+ {
+ "epoch": 29.9198606271777,
+ "grad_norm": 0.0640307068824768,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 1.9258478879928589,
+ "step": 2154
+ },
+ {
+ "epoch": 29.933797909407666,
+ "grad_norm": 0.06517527252435684,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 1.9236748218536377,
+ "step": 2155
+ },
+ {
+ "epoch": 29.94773519163763,
+ "grad_norm": 0.06373083591461182,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 1.9360520839691162,
+ "step": 2156
+ },
+ {
+ "epoch": 29.961672473867594,
+ "grad_norm": 0.06771238893270493,
+ "learning_rate": 0.00022244633283095,
+ "loss": 1.947780966758728,
+ "step": 2157
+ },
+ {
+ "epoch": 29.975609756097562,
+ "grad_norm": 0.06298559159040451,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 1.9223754405975342,
+ "step": 2158
+ },
+ {
+ "epoch": 29.989547038327526,
+ "grad_norm": 0.06455206871032715,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 1.9330079555511475,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.0522056445479393,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 1.4463932514190674,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6358165740966797,
+ "eval_runtime": 40.2782,
+ "eval_samples_per_second": 60.628,
+ "eval_steps_per_second": 0.497,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013937282229964,
+ "grad_norm": 0.06026934087276459,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 1.9035022258758545,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027874564459932,
+ "grad_norm": 0.06600011885166168,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 1.8965950012207031,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041811846689896,
+ "grad_norm": 0.060655586421489716,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 1.8954249620437622,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05574912891986,
+ "grad_norm": 0.0636422410607338,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 1.8951303958892822,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069686411149824,
+ "grad_norm": 0.05972317233681679,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 1.910957932472229,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083623693379792,
+ "grad_norm": 0.060592107474803925,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 1.9128470420837402,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097560975609756,
+ "grad_norm": 0.05888872966170311,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 1.897517204284668,
+ "step": 2167
+ },
+ {
+ "epoch": 30.11149825783972,
+ "grad_norm": 0.062287602573633194,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 1.9048705101013184,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125435540069688,
+ "grad_norm": 0.05919402465224266,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 1.901914119720459,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139372822299652,
+ "grad_norm": 0.06069913133978844,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 1.9066851139068604,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153310104529616,
+ "grad_norm": 0.06046717241406441,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 1.8969058990478516,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16724738675958,
+ "grad_norm": 0.059624481946229935,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 1.8908498287200928,
+ "step": 2172
+ },
+ {
+ "epoch": 30.181184668989548,
+ "grad_norm": 0.06100156158208847,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 1.909839391708374,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195121951219512,
+ "grad_norm": 0.05801135674118996,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 1.9057464599609375,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209059233449477,
+ "grad_norm": 0.06198010593652725,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 1.9051663875579834,
+ "step": 2175
+ },
+ {
+ "epoch": 30.222996515679444,
+ "grad_norm": 0.06107710674405098,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 1.904172658920288,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23693379790941,
+ "grad_norm": 0.060085635632276535,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 1.8970673084259033,
+ "step": 2177
+ },
+ {
+ "epoch": 30.250871080139373,
+ "grad_norm": 0.05755467712879181,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 1.9092798233032227,
+ "step": 2178
+ },
+ {
+ "epoch": 30.264808362369337,
+ "grad_norm": 0.05827959626913071,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 1.9012596607208252,
+ "step": 2179
+ },
+ {
+ "epoch": 30.278745644599304,
+ "grad_norm": 0.0570731982588768,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 1.9027369022369385,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29268292682927,
+ "grad_norm": 0.05898377671837807,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 1.9284251928329468,
+ "step": 2181
+ },
+ {
+ "epoch": 30.306620209059233,
+ "grad_norm": 0.06119367107748985,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 1.8992201089859009,
+ "step": 2182
+ },
+ {
+ "epoch": 30.320557491289197,
+ "grad_norm": 0.05790785327553749,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 1.9106863737106323,
+ "step": 2183
+ },
+ {
+ "epoch": 30.334494773519165,
+ "grad_norm": 0.06101157143712044,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 1.9129860401153564,
+ "step": 2184
+ },
+ {
+ "epoch": 30.34843205574913,
+ "grad_norm": 0.062228068709373474,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 1.9075040817260742,
+ "step": 2185
+ },
+ {
+ "epoch": 30.362369337979093,
+ "grad_norm": 0.057724758982658386,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 1.9021648168563843,
+ "step": 2186
+ },
+ {
+ "epoch": 30.37630662020906,
+ "grad_norm": 0.05886731296777725,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 1.8955028057098389,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390243902439025,
+ "grad_norm": 0.0613555908203125,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 1.8902294635772705,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40418118466899,
+ "grad_norm": 0.059424031525850296,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 1.9234192371368408,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418118466898953,
+ "grad_norm": 0.059314578771591187,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 1.9120237827301025,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43205574912892,
+ "grad_norm": 0.059916652739048004,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 1.9118289947509766,
+ "step": 2191
+ },
+ {
+ "epoch": 30.445993031358885,
+ "grad_norm": 0.05817565694451332,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 1.9012054204940796,
+ "step": 2192
+ },
+ {
+ "epoch": 30.45993031358885,
+ "grad_norm": 0.061957065016031265,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 1.9110759496688843,
+ "step": 2193
+ },
+ {
+ "epoch": 30.473867595818817,
+ "grad_norm": 0.05899856239557266,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 1.9285509586334229,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48780487804878,
+ "grad_norm": 0.06001519039273262,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 1.9093801975250244,
+ "step": 2195
+ },
+ {
+ "epoch": 30.501742160278745,
+ "grad_norm": 0.060825154185295105,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 1.892765760421753,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51567944250871,
+ "grad_norm": 0.06136799231171608,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 1.9276010990142822,
+ "step": 2197
+ },
+ {
+ "epoch": 30.529616724738677,
+ "grad_norm": 0.05773079767823219,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 1.9083237648010254,
+ "step": 2198
+ },
+ {
+ "epoch": 30.54355400696864,
+ "grad_norm": 0.06120939180254936,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 1.8891932964324951,
+ "step": 2199
+ },
+ {
+ "epoch": 30.557491289198605,
+ "grad_norm": 0.06097486987709999,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 1.9084399938583374,
+ "step": 2200
+ },
+ {
+ "epoch": 30.571428571428573,
+ "grad_norm": 0.05973029136657715,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 1.918110966682434,
+ "step": 2201
+ },
+ {
+ "epoch": 30.585365853658537,
+ "grad_norm": 0.06335631012916565,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 1.9174532890319824,
+ "step": 2202
+ },
+ {
+ "epoch": 30.5993031358885,
+ "grad_norm": 0.05974910035729408,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 1.9107253551483154,
+ "step": 2203
+ },
+ {
+ "epoch": 30.613240418118465,
+ "grad_norm": 0.06281539052724838,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 1.9007208347320557,
+ "step": 2204
+ },
+ {
+ "epoch": 30.627177700348433,
+ "grad_norm": 0.062120143324136734,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 1.911742925643921,
+ "step": 2205
+ },
+ {
+ "epoch": 30.641114982578397,
+ "grad_norm": 0.06305570155382156,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 1.9121983051300049,
+ "step": 2206
+ },
+ {
+ "epoch": 30.65505226480836,
+ "grad_norm": 0.06112781539559364,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 1.908568263053894,
+ "step": 2207
+ },
+ {
+ "epoch": 30.66898954703833,
+ "grad_norm": 0.06050721928477287,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 1.9018139839172363,
+ "step": 2208
+ },
+ {
+ "epoch": 30.682926829268293,
+ "grad_norm": 0.05980826914310455,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 1.9104691743850708,
+ "step": 2209
+ },
+ {
+ "epoch": 30.696864111498257,
+ "grad_norm": 0.06286003440618515,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 1.899999737739563,
+ "step": 2210
+ },
+ {
+ "epoch": 30.71080139372822,
+ "grad_norm": 0.06179053336381912,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 1.8993868827819824,
+ "step": 2211
+ },
+ {
+ "epoch": 30.72473867595819,
+ "grad_norm": 0.06006763502955437,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 1.9067524671554565,
+ "step": 2212
+ },
+ {
+ "epoch": 30.738675958188153,
+ "grad_norm": 0.06104302033782005,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 1.9074416160583496,
+ "step": 2213
+ },
+ {
+ "epoch": 30.752613240418118,
+ "grad_norm": 0.059952687472105026,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 1.9049818515777588,
+ "step": 2214
+ },
+ {
+ "epoch": 30.766550522648085,
+ "grad_norm": 0.06469111144542694,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 1.9107921123504639,
+ "step": 2215
+ },
+ {
+ "epoch": 30.78048780487805,
+ "grad_norm": 0.06027381867170334,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 1.920630693435669,
+ "step": 2216
+ },
+ {
+ "epoch": 30.794425087108014,
+ "grad_norm": 0.06066177412867546,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 1.9131195545196533,
+ "step": 2217
+ },
+ {
+ "epoch": 30.808362369337978,
+ "grad_norm": 0.06533747911453247,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 1.9172463417053223,
+ "step": 2218
+ },
+ {
+ "epoch": 30.822299651567945,
+ "grad_norm": 0.061867427080869675,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 1.9100360870361328,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83623693379791,
+ "grad_norm": 0.06394216418266296,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 1.9218776226043701,
+ "step": 2220
+ },
+ {
+ "epoch": 30.850174216027874,
+ "grad_norm": 0.061471469700336456,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 1.901368498802185,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86411149825784,
+ "grad_norm": 0.06380587071180344,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 1.9187582731246948,
+ "step": 2222
+ },
+ {
+ "epoch": 30.878048780487806,
+ "grad_norm": 0.06244281679391861,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 1.9082577228546143,
+ "step": 2223
+ },
+ {
+ "epoch": 30.89198606271777,
+ "grad_norm": 0.06435921788215637,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 1.924039363861084,
+ "step": 2224
+ },
+ {
+ "epoch": 30.905923344947734,
+ "grad_norm": 0.06331372261047363,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 1.9083380699157715,
+ "step": 2225
+ },
+ {
+ "epoch": 30.9198606271777,
+ "grad_norm": 0.06393173336982727,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 1.9132580757141113,
+ "step": 2226
+ },
+ {
+ "epoch": 30.933797909407666,
+ "grad_norm": 0.06500784307718277,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 1.9109617471694946,
+ "step": 2227
+ },
+ {
+ "epoch": 30.94773519163763,
+ "grad_norm": 0.06389841437339783,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 1.926948070526123,
+ "step": 2228
+ },
+ {
+ "epoch": 30.961672473867594,
+ "grad_norm": 0.06343282759189606,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 1.9304101467132568,
+ "step": 2229
+ },
+ {
+ "epoch": 30.975609756097562,
+ "grad_norm": 0.06003900617361069,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 1.9066693782806396,
+ "step": 2230
+ },
+ {
+ "epoch": 30.989547038327526,
+ "grad_norm": 0.06355264037847519,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 1.9183707237243652,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0492861270904541,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 1.4322078227996826,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6370227932929993,
+ "eval_runtime": 40.5546,
+ "eval_samples_per_second": 60.215,
+ "eval_steps_per_second": 0.493,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013937282229964,
+ "grad_norm": 0.057478148490190506,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 1.887700080871582,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027874564459932,
+ "grad_norm": 0.06220356002449989,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 1.8875237703323364,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041811846689896,
+ "grad_norm": 0.05686607584357262,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 1.897495150566101,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05574912891986,
+ "grad_norm": 0.06147320941090584,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 1.8913912773132324,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069686411149824,
+ "grad_norm": 0.06075664609670639,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 1.8823286294937134,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083623693379792,
+ "grad_norm": 0.058347299695014954,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 1.8890151977539062,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097560975609756,
+ "grad_norm": 0.06388939172029495,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 1.8857264518737793,
+ "step": 2239
+ },
+ {
+ "epoch": 31.11149825783972,
+ "grad_norm": 0.05901556834578514,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 1.8779339790344238,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125435540069688,
+ "grad_norm": 0.06094517558813095,
+ "learning_rate": 0.000201141724176723,
+ "loss": 1.8811697959899902,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139372822299652,
+ "grad_norm": 0.060666318982839584,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 1.8911728858947754,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153310104529616,
+ "grad_norm": 0.05608760938048363,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 1.8804845809936523,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16724738675958,
+ "grad_norm": 0.06151990965008736,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 1.8901598453521729,
+ "step": 2244
+ },
+ {
+ "epoch": 31.181184668989548,
+ "grad_norm": 0.05805948004126549,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 1.8834660053253174,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195121951219512,
+ "grad_norm": 0.05967888608574867,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 1.8807945251464844,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209059233449477,
+ "grad_norm": 0.06295159459114075,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 1.8946118354797363,
+ "step": 2247
+ },
+ {
+ "epoch": 31.222996515679444,
+ "grad_norm": 0.057971853762865067,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 1.889994502067566,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23693379790941,
+ "grad_norm": 0.05941799655556679,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 1.884669542312622,
+ "step": 2249
+ },
+ {
+ "epoch": 31.250871080139373,
+ "grad_norm": 0.06173485890030861,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 1.8727672100067139,
+ "step": 2250
+ },
+ {
+ "epoch": 31.264808362369337,
+ "grad_norm": 0.05898866057395935,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 1.8795123100280762,
+ "step": 2251
+ },
+ {
+ "epoch": 31.278745644599304,
+ "grad_norm": 0.06072458624839783,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 1.8846805095672607,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29268292682927,
+ "grad_norm": 0.060174115002155304,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 1.8993717432022095,
+ "step": 2253
+ },
+ {
+ "epoch": 31.306620209059233,
+ "grad_norm": 0.061605289578437805,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 1.8877589702606201,
+ "step": 2254
+ },
+ {
+ "epoch": 31.320557491289197,
+ "grad_norm": 0.06246590241789818,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 1.8974242210388184,
+ "step": 2255
+ },
+ {
+ "epoch": 31.334494773519165,
+ "grad_norm": 0.060071494430303574,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 1.881445288658142,
+ "step": 2256
+ },
+ {
+ "epoch": 31.34843205574913,
+ "grad_norm": 0.060360442847013474,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 1.8938015699386597,
+ "step": 2257
+ },
+ {
+ "epoch": 31.362369337979093,
+ "grad_norm": 0.06192562356591225,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 1.8961126804351807,
+ "step": 2258
+ },
+ {
+ "epoch": 31.37630662020906,
+ "grad_norm": 0.05872832611203194,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 1.8974393606185913,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390243902439025,
+ "grad_norm": 0.05980943515896797,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 1.8886656761169434,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40418118466899,
+ "grad_norm": 0.0593055821955204,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 1.8888055086135864,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418118466898953,
+ "grad_norm": 0.059456177055835724,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 1.9045109748840332,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43205574912892,
+ "grad_norm": 0.06260113418102264,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 1.9062271118164062,
+ "step": 2263
+ },
+ {
+ "epoch": 31.445993031358885,
+ "grad_norm": 0.05866454914212227,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 1.8920798301696777,
+ "step": 2264
+ },
+ {
+ "epoch": 31.45993031358885,
+ "grad_norm": 0.060281749814748764,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 1.9048649072647095,
+ "step": 2265
+ },
+ {
+ "epoch": 31.473867595818817,
+ "grad_norm": 0.060433730483055115,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 1.8905510902404785,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48780487804878,
+ "grad_norm": 0.06074376404285431,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 1.890744686126709,
+ "step": 2267
+ },
+ {
+ "epoch": 31.501742160278745,
+ "grad_norm": 0.060131702572107315,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 1.893791913986206,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51567944250871,
+ "grad_norm": 0.05804680287837982,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 1.8838505744934082,
+ "step": 2269
+ },
+ {
+ "epoch": 31.529616724738677,
+ "grad_norm": 0.05933694168925285,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 1.897843360900879,
+ "step": 2270
+ },
+ {
+ "epoch": 31.54355400696864,
+ "grad_norm": 0.058537356555461884,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 1.8826205730438232,
+ "step": 2271
+ },
+ {
+ "epoch": 31.557491289198605,
+ "grad_norm": 0.060081496834754944,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 1.889431357383728,
+ "step": 2272
+ },
+ {
+ "epoch": 31.571428571428573,
+ "grad_norm": 0.05904357507824898,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 1.8965020179748535,
+ "step": 2273
+ },
+ {
+ "epoch": 31.585365853658537,
+ "grad_norm": 0.05924354866147041,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 1.894204020500183,
+ "step": 2274
+ },
+ {
+ "epoch": 31.5993031358885,
+ "grad_norm": 0.059307269752025604,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 1.890209436416626,
+ "step": 2275
+ },
+ {
+ "epoch": 31.613240418118465,
+ "grad_norm": 0.05969221144914627,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 1.8865056037902832,
+ "step": 2276
+ },
+ {
+ "epoch": 31.627177700348433,
+ "grad_norm": 0.060738109052181244,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 1.8825939893722534,
+ "step": 2277
+ },
+ {
+ "epoch": 31.641114982578397,
+ "grad_norm": 0.06001656875014305,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 1.8988475799560547,
+ "step": 2278
+ },
+ {
+ "epoch": 31.65505226480836,
+ "grad_norm": 0.061478059738874435,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 1.9071214199066162,
+ "step": 2279
+ },
+ {
+ "epoch": 31.66898954703833,
+ "grad_norm": 0.06089174002408981,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 1.8998515605926514,
+ "step": 2280
+ },
+ {
+ "epoch": 31.682926829268293,
+ "grad_norm": 0.05940974876284599,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 1.8907361030578613,
+ "step": 2281
+ },
+ {
+ "epoch": 31.696864111498257,
+ "grad_norm": 0.06015758961439133,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 1.9010521173477173,
+ "step": 2282
+ },
+ {
+ "epoch": 31.71080139372822,
+ "grad_norm": 0.06062757968902588,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 1.899493932723999,
+ "step": 2283
+ },
+ {
+ "epoch": 31.72473867595819,
+ "grad_norm": 0.060489553958177567,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 1.9095027446746826,
+ "step": 2284
+ },
+ {
+ "epoch": 31.738675958188153,
+ "grad_norm": 0.05949694290757179,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 1.900972604751587,
+ "step": 2285
+ },
+ {
+ "epoch": 31.752613240418118,
+ "grad_norm": 0.06269118189811707,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 1.891621708869934,
+ "step": 2286
+ },
+ {
+ "epoch": 31.766550522648085,
+ "grad_norm": 0.06033318489789963,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 1.8954113721847534,
+ "step": 2287
+ },
+ {
+ "epoch": 31.78048780487805,
+ "grad_norm": 0.0604838989675045,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 1.9047741889953613,
+ "step": 2288
+ },
+ {
+ "epoch": 31.794425087108014,
+ "grad_norm": 0.060087885707616806,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 1.8912427425384521,
+ "step": 2289
+ },
+ {
+ "epoch": 31.808362369337978,
+ "grad_norm": 0.059592097997665405,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 1.8874431848526,
+ "step": 2290
+ },
+ {
+ "epoch": 31.822299651567945,
+ "grad_norm": 0.06219378113746643,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 1.8843477964401245,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83623693379791,
+ "grad_norm": 0.06010756641626358,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 1.892686367034912,
+ "step": 2292
+ },
+ {
+ "epoch": 31.850174216027874,
+ "grad_norm": 0.061404358595609665,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 1.907541275024414,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86411149825784,
+ "grad_norm": 0.060503825545310974,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 1.891597032546997,
+ "step": 2294
+ },
+ {
+ "epoch": 31.878048780487806,
+ "grad_norm": 0.06197115406394005,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 1.8968595266342163,
+ "step": 2295
+ },
+ {
+ "epoch": 31.89198606271777,
+ "grad_norm": 0.06244310364127159,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 1.903473138809204,
+ "step": 2296
+ },
+ {
+ "epoch": 31.905923344947734,
+ "grad_norm": 0.060648202896118164,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 1.9044740200042725,
+ "step": 2297
+ },
+ {
+ "epoch": 31.9198606271777,
+ "grad_norm": 0.061299171298742294,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 1.8890066146850586,
+ "step": 2298
+ },
+ {
+ "epoch": 31.933797909407666,
+ "grad_norm": 0.06204016134142876,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 1.912150502204895,
+ "step": 2299
+ },
+ {
+ "epoch": 31.94773519163763,
+ "grad_norm": 0.06292745471000671,
+ "learning_rate": 0.000186516746349841,
+ "loss": 1.9133987426757812,
+ "step": 2300
+ },
+ {
+ "epoch": 31.961672473867594,
+ "grad_norm": 0.0633154958486557,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 1.9106671810150146,
+ "step": 2301
+ },
+ {
+ "epoch": 31.975609756097562,
+ "grad_norm": 0.06269625574350357,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 1.9090297222137451,
+ "step": 2302
+ },
+ {
+ "epoch": 31.989547038327526,
+ "grad_norm": 0.06060625612735748,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 1.8874908685684204,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.05050232633948326,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 1.4178190231323242,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6386013627052307,
+ "eval_runtime": 40.9718,
+ "eval_samples_per_second": 59.602,
+ "eval_steps_per_second": 0.488,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01393728222997,
+ "grad_norm": 0.05701690912246704,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 1.8772780895233154,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02787456445993,
+ "grad_norm": 0.05597634240984917,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 1.8772567510604858,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041811846689896,
+ "grad_norm": 0.05631265789270401,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 1.8660085201263428,
+ "step": 2307
+ },
+ {
+ "epoch": 32.055749128919864,
+ "grad_norm": 0.05604572966694832,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 1.876168966293335,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069686411149824,
+ "grad_norm": 0.056123036891222,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 1.8643467426300049,
+ "step": 2309
+ },
+ {
+ "epoch": 32.08362369337979,
+ "grad_norm": 0.05791968107223511,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 1.866549015045166,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09756097560975,
+ "grad_norm": 0.05521981045603752,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 1.8852781057357788,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11149825783972,
+ "grad_norm": 0.055954866111278534,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 1.8603641986846924,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12543554006969,
+ "grad_norm": 0.05931234359741211,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 1.8619375228881836,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13937282229965,
+ "grad_norm": 0.05743015930056572,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 1.8824023008346558,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153310104529616,
+ "grad_norm": 0.056393202394247055,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 1.8722079992294312,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167247386759584,
+ "grad_norm": 0.05671187862753868,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 1.8824248313903809,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181184668989545,
+ "grad_norm": 0.05667170509696007,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 1.882739782333374,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19512195121951,
+ "grad_norm": 0.05515167862176895,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 1.8555771112442017,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20905923344948,
+ "grad_norm": 0.05649147555232048,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 1.8673419952392578,
+ "step": 2319
+ },
+ {
+ "epoch": 32.22299651567944,
+ "grad_norm": 0.058623913675546646,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 1.8744664192199707,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23693379790941,
+ "grad_norm": 0.055131133645772934,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 1.8781135082244873,
+ "step": 2321
+ },
+ {
+ "epoch": 32.250871080139376,
+ "grad_norm": 0.06008831411600113,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 1.8844444751739502,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26480836236934,
+ "grad_norm": 0.05755964294075966,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 1.8790829181671143,
+ "step": 2323
+ },
+ {
+ "epoch": 32.278745644599304,
+ "grad_norm": 0.055803194642066956,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 1.8642692565917969,
+ "step": 2324
+ },
+ {
+ "epoch": 32.292682926829265,
+ "grad_norm": 0.05710228905081749,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 1.8659579753875732,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30662020905923,
+ "grad_norm": 0.05771065130829811,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 1.8812355995178223,
+ "step": 2326
+ },
+ {
+ "epoch": 32.3205574912892,
+ "grad_norm": 0.05642757564783096,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 1.8610783815383911,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33449477351916,
+ "grad_norm": 0.05687689036130905,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 1.875876545906067,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34843205574913,
+ "grad_norm": 0.057835984975099564,
+ "learning_rate": 0.000179445406945268,
+ "loss": 1.8674073219299316,
+ "step": 2329
+ },
+ {
+ "epoch": 32.362369337979096,
+ "grad_norm": 0.05800129100680351,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 1.8708086013793945,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37630662020906,
+ "grad_norm": 0.05744084715843201,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 1.8869245052337646,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390243902439025,
+ "grad_norm": 0.056703388690948486,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 1.8688050508499146,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40418118466899,
+ "grad_norm": 0.06032602861523628,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 1.873382568359375,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41811846689895,
+ "grad_norm": 0.05810776352882385,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 1.8624725341796875,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43205574912892,
+ "grad_norm": 0.060990724712610245,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 1.8792011737823486,
+ "step": 2335
+ },
+ {
+ "epoch": 32.44599303135889,
+ "grad_norm": 0.05975554883480072,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 1.8793275356292725,
+ "step": 2336
+ },
+ {
+ "epoch": 32.45993031358885,
+ "grad_norm": 0.05772533640265465,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 1.889489769935608,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47386759581882,
+ "grad_norm": 0.05920211225748062,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 1.8647074699401855,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48780487804878,
+ "grad_norm": 0.059745173901319504,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 1.8649309873580933,
+ "step": 2339
+ },
+ {
+ "epoch": 32.501742160278745,
+ "grad_norm": 0.05933282524347305,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 1.8936223983764648,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51567944250871,
+ "grad_norm": 0.062407393008470535,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 1.8831725120544434,
+ "step": 2341
+ },
+ {
+ "epoch": 32.52961672473867,
+ "grad_norm": 0.061185337603092194,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 1.8766106367111206,
+ "step": 2342
+ },
+ {
+ "epoch": 32.54355400696864,
+ "grad_norm": 0.06082421913743019,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 1.880276083946228,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55749128919861,
+ "grad_norm": 0.05900120362639427,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 1.8696876764297485,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57142857142857,
+ "grad_norm": 0.06146271899342537,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 1.864588737487793,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58536585365854,
+ "grad_norm": 0.05976773053407669,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 1.8771488666534424,
+ "step": 2346
+ },
+ {
+ "epoch": 32.599303135888505,
+ "grad_norm": 0.05808904021978378,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 1.890838623046875,
+ "step": 2347
+ },
+ {
+ "epoch": 32.613240418118465,
+ "grad_norm": 0.06149672716856003,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 1.8787143230438232,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62717770034843,
+ "grad_norm": 0.059028271585702896,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 1.8875718116760254,
+ "step": 2349
+ },
+ {
+ "epoch": 32.641114982578394,
+ "grad_norm": 0.062163304537534714,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 1.86748206615448,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65505226480836,
+ "grad_norm": 0.06221195310354233,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 1.8742306232452393,
+ "step": 2351
+ },
+ {
+ "epoch": 32.66898954703833,
+ "grad_norm": 0.060461003333330154,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 1.885633945465088,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68292682926829,
+ "grad_norm": 0.06130106747150421,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 1.8827314376831055,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69686411149826,
+ "grad_norm": 0.06044832617044449,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 1.8724548816680908,
+ "step": 2354
+ },
+ {
+ "epoch": 32.710801393728225,
+ "grad_norm": 0.06029713526368141,
+ "learning_rate": 0.000173176617304673,
+ "loss": 1.8745346069335938,
+ "step": 2355
+ },
+ {
+ "epoch": 32.724738675958186,
+ "grad_norm": 0.06287439167499542,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 1.8879387378692627,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73867595818815,
+ "grad_norm": 0.059636496007442474,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 1.882830262184143,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75261324041812,
+ "grad_norm": 0.06590992957353592,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 1.882242202758789,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76655052264808,
+ "grad_norm": 0.06283026933670044,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 1.8864011764526367,
+ "step": 2359
+ },
+ {
+ "epoch": 32.78048780487805,
+ "grad_norm": 0.061717696487903595,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 1.8815748691558838,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79442508710802,
+ "grad_norm": 0.06423293799161911,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 1.8824026584625244,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80836236933798,
+ "grad_norm": 0.06351540237665176,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 1.9000264406204224,
+ "step": 2362
+ },
+ {
+ "epoch": 32.822299651567945,
+ "grad_norm": 0.060434527695178986,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 1.8875732421875,
+ "step": 2363
+ },
+ {
+ "epoch": 32.836236933797906,
+ "grad_norm": 0.062305059283971786,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 1.8695244789123535,
+ "step": 2364
+ },
+ {
+ "epoch": 32.850174216027874,
+ "grad_norm": 0.05981886014342308,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 1.8885269165039062,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86411149825784,
+ "grad_norm": 0.0642906129360199,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 1.8818528652191162,
+ "step": 2366
+ },
+ {
+ "epoch": 32.8780487804878,
+ "grad_norm": 0.06316667050123215,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 1.8994319438934326,
+ "step": 2367
+ },
+ {
+ "epoch": 32.89198606271777,
+ "grad_norm": 0.058260831981897354,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 1.8908791542053223,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90592334494774,
+ "grad_norm": 0.062237389385700226,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 1.8753025531768799,
+ "step": 2369
+ },
+ {
+ "epoch": 32.9198606271777,
+ "grad_norm": 0.05972493812441826,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 1.899425745010376,
+ "step": 2370
+ },
+ {
+ "epoch": 32.933797909407666,
+ "grad_norm": 0.058846332132816315,
+ "learning_rate": 0.00016935382741164,
+ "loss": 1.8934335708618164,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94773519163763,
+ "grad_norm": 0.06333424896001816,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 1.8951395750045776,
+ "step": 2372
+ },
+ {
+ "epoch": 32.961672473867594,
+ "grad_norm": 0.06063693389296532,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 1.881688117980957,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97560975609756,
+ "grad_norm": 0.060873813927173615,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 1.8836770057678223,
+ "step": 2374
+ },
+ {
+ "epoch": 32.98954703832753,
+ "grad_norm": 0.060378748923540115,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 1.886624813079834,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.04906065762042999,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 1.4146361351013184,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6401442885398865,
+ "eval_runtime": 41.0841,
+ "eval_samples_per_second": 59.439,
+ "eval_steps_per_second": 0.487,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01393728222997,
+ "grad_norm": 0.05557359382510185,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 1.8605964183807373,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02787456445993,
+ "grad_norm": 0.055585991591215134,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 1.8456363677978516,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041811846689896,
+ "grad_norm": 0.05610295757651329,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 1.847888469696045,
+ "step": 2379
+ },
+ {
+ "epoch": 33.055749128919864,
+ "grad_norm": 0.05532662197947502,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 1.850597620010376,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069686411149824,
+ "grad_norm": 0.05585319921374321,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 1.8694689273834229,
+ "step": 2381
+ },
+ {
+ "epoch": 33.08362369337979,
+ "grad_norm": 0.05658521130681038,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 1.8533908128738403,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09756097560975,
+ "grad_norm": 0.0558413490653038,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 1.8735582828521729,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11149825783972,
+ "grad_norm": 0.05637028068304062,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 1.8585509061813354,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12543554006969,
+ "grad_norm": 0.055854618549346924,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 1.8539409637451172,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13937282229965,
+ "grad_norm": 0.055990077555179596,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 1.85886812210083,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153310104529616,
+ "grad_norm": 0.05781752988696098,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 1.8603525161743164,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167247386759584,
+ "grad_norm": 0.055855557322502136,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 1.861933946609497,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181184668989545,
+ "grad_norm": 0.05729927122592926,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 1.86094069480896,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19512195121951,
+ "grad_norm": 0.056495532393455505,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 1.848013997077942,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20905923344948,
+ "grad_norm": 0.0571073479950428,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 1.8558173179626465,
+ "step": 2391
+ },
+ {
+ "epoch": 33.22299651567944,
+ "grad_norm": 0.056066226214170456,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 1.8723382949829102,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23693379790941,
+ "grad_norm": 0.05570382624864578,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 1.864349365234375,
+ "step": 2393
+ },
+ {
+ "epoch": 33.250871080139376,
+ "grad_norm": 0.054912492632865906,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 1.8678479194641113,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26480836236934,
+ "grad_norm": 0.05622183158993721,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 1.8648667335510254,
+ "step": 2395
+ },
+ {
+ "epoch": 33.278745644599304,
+ "grad_norm": 0.05719740688800812,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 1.850623607635498,
+ "step": 2396
+ },
+ {
+ "epoch": 33.292682926829265,
+ "grad_norm": 0.056594349443912506,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 1.8707754611968994,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30662020905923,
+ "grad_norm": 0.05596613511443138,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 1.8580527305603027,
+ "step": 2398
+ },
+ {
+ "epoch": 33.3205574912892,
+ "grad_norm": 0.05675627663731575,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 1.8687434196472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33449477351916,
+ "grad_norm": 0.05572087690234184,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 1.8603391647338867,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34843205574913,
+ "grad_norm": 0.05837463587522507,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 1.8638341426849365,
+ "step": 2401
+ },
+ {
+ "epoch": 33.362369337979096,
+ "grad_norm": 0.056136030703783035,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 1.853350281715393,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37630662020906,
+ "grad_norm": 0.0555158406496048,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 1.8775548934936523,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390243902439025,
+ "grad_norm": 0.0595950111746788,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 1.8578253984451294,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40418118466899,
+ "grad_norm": 0.05565760284662247,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 1.8508214950561523,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41811846689895,
+ "grad_norm": 0.058832235634326935,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 1.8614635467529297,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43205574912892,
+ "grad_norm": 0.0585586316883564,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 1.8617370128631592,
+ "step": 2407
+ },
+ {
+ "epoch": 33.44599303135889,
+ "grad_norm": 0.05951899662613869,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 1.8648651838302612,
+ "step": 2408
+ },
+ {
+ "epoch": 33.45993031358885,
+ "grad_norm": 0.057204652577638626,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 1.8577271699905396,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47386759581882,
+ "grad_norm": 0.05844349041581154,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 1.8560724258422852,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48780487804878,
+ "grad_norm": 0.05761651322245598,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 1.863309383392334,
+ "step": 2411
+ },
+ {
+ "epoch": 33.501742160278745,
+ "grad_norm": 0.05818328261375427,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 1.8798553943634033,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51567944250871,
+ "grad_norm": 0.05746658146381378,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 1.8628616333007812,
+ "step": 2413
+ },
+ {
+ "epoch": 33.52961672473867,
+ "grad_norm": 0.057656314224004745,
+ "learning_rate": 0.000159218843594243,
+ "loss": 1.8699405193328857,
+ "step": 2414
+ },
+ {
+ "epoch": 33.54355400696864,
+ "grad_norm": 0.0586664155125618,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 1.8481391668319702,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55749128919861,
+ "grad_norm": 0.05769997090101242,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 1.8612866401672363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57142857142857,
+ "grad_norm": 0.0598456896841526,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 1.8580410480499268,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58536585365854,
+ "grad_norm": 0.05761442333459854,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 1.874891996383667,
+ "step": 2418
+ },
+ {
+ "epoch": 33.599303135888505,
+ "grad_norm": 0.06176742538809776,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 1.8660908937454224,
+ "step": 2419
+ },
+ {
+ "epoch": 33.613240418118465,
+ "grad_norm": 0.0588943213224411,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 1.8718886375427246,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62717770034843,
+ "grad_norm": 0.06092666834592819,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 1.8634107112884521,
+ "step": 2421
+ },
+ {
+ "epoch": 33.641114982578394,
+ "grad_norm": 0.05774684250354767,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 1.868361234664917,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65505226480836,
+ "grad_norm": 0.05932781100273132,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 1.867889642715454,
+ "step": 2423
+ },
+ {
+ "epoch": 33.66898954703833,
+ "grad_norm": 0.058603718876838684,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 1.8793346881866455,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68292682926829,
+ "grad_norm": 0.05870316922664642,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 1.8717231750488281,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69686411149826,
+ "grad_norm": 0.05805273354053497,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 1.8479297161102295,
+ "step": 2426
+ },
+ {
+ "epoch": 33.710801393728225,
+ "grad_norm": 0.059018444269895554,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 1.8711276054382324,
+ "step": 2427
+ },
+ {
+ "epoch": 33.724738675958186,
+ "grad_norm": 0.05927985534071922,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 1.8653428554534912,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73867595818815,
+ "grad_norm": 0.06085766851902008,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 1.8724851608276367,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75261324041812,
+ "grad_norm": 0.05974132940173149,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 1.8459011316299438,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76655052264808,
+ "grad_norm": 0.05851755291223526,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 1.863109827041626,
+ "step": 2431
+ },
+ {
+ "epoch": 33.78048780487805,
+ "grad_norm": 0.06079421192407608,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 1.8687033653259277,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79442508710802,
+ "grad_norm": 0.057930078357458115,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 1.8653396368026733,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80836236933798,
+ "grad_norm": 0.059440065175294876,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 1.8569486141204834,
+ "step": 2434
+ },
+ {
+ "epoch": 33.822299651567945,
+ "grad_norm": 0.05835104361176491,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 1.8828535079956055,
+ "step": 2435
+ },
+ {
+ "epoch": 33.836236933797906,
+ "grad_norm": 0.06108623370528221,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 1.8749947547912598,
+ "step": 2436
+ },
+ {
+ "epoch": 33.850174216027874,
+ "grad_norm": 0.059921979904174805,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 1.8684660196304321,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86411149825784,
+ "grad_norm": 0.059964295476675034,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 1.862084150314331,
+ "step": 2438
+ },
+ {
+ "epoch": 33.8780487804878,
+ "grad_norm": 0.05888533219695091,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 1.85898756980896,
+ "step": 2439
+ },
+ {
+ "epoch": 33.89198606271777,
+ "grad_norm": 0.058946020901203156,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 1.872513771057129,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90592334494774,
+ "grad_norm": 0.05947734788060188,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 1.8610495328903198,
+ "step": 2441
+ },
+ {
+ "epoch": 33.9198606271777,
+ "grad_norm": 0.05859588086605072,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 1.8842005729675293,
+ "step": 2442
+ },
+ {
+ "epoch": 33.933797909407666,
+ "grad_norm": 0.05818016827106476,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 1.8715083599090576,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94773519163763,
+ "grad_norm": 0.05960588902235031,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 1.881840467453003,
+ "step": 2444
+ },
+ {
+ "epoch": 33.961672473867594,
+ "grad_norm": 0.059536516666412354,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 1.8734736442565918,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97560975609756,
+ "grad_norm": 0.05922378972172737,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 1.8569964170455933,
+ "step": 2446
+ },
+ {
+ "epoch": 33.98954703832753,
+ "grad_norm": 0.058573752641677856,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 1.8710730075836182,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.047325484454631805,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 1.4022152423858643,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6414204835891724,
+ "eval_runtime": 39.9489,
+ "eval_samples_per_second": 61.128,
+ "eval_steps_per_second": 0.501,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01393728222997,
+ "grad_norm": 0.05358244478702545,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 1.8384610414505005,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02787456445993,
+ "grad_norm": 0.0558239221572876,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 1.8538508415222168,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041811846689896,
+ "grad_norm": 0.05390622094273567,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 1.8460195064544678,
+ "step": 2451
+ },
+ {
+ "epoch": 34.055749128919864,
+ "grad_norm": 0.056224383413791656,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 1.8416059017181396,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069686411149824,
+ "grad_norm": 0.053925771266222,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 1.8532803058624268,
+ "step": 2453
+ },
+ {
+ "epoch": 34.08362369337979,
+ "grad_norm": 0.05750057101249695,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 1.847919225692749,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09756097560975,
+ "grad_norm": 0.05499042198061943,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 1.8450297117233276,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11149825783972,
+ "grad_norm": 0.0544770173728466,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 1.8482413291931152,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12543554006969,
+ "grad_norm": 0.05656695365905762,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 1.860766053199768,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13937282229965,
+ "grad_norm": 0.054617058485746384,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 1.8398511409759521,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153310104529616,
+ "grad_norm": 0.05470099300146103,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 1.8501665592193604,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167247386759584,
+ "grad_norm": 0.05437156930565834,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 1.8514611721038818,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181184668989545,
+ "grad_norm": 0.05481892451643944,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 1.8424533605575562,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19512195121951,
+ "grad_norm": 0.055016402155160904,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 1.823396921157837,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20905923344948,
+ "grad_norm": 0.05608086287975311,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 1.8525217771530151,
+ "step": 2463
+ },
+ {
+ "epoch": 34.22299651567944,
+ "grad_norm": 0.055000413209199905,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 1.8570690155029297,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23693379790941,
+ "grad_norm": 0.05437581241130829,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 1.8471980094909668,
+ "step": 2465
+ },
+ {
+ "epoch": 34.250871080139376,
+ "grad_norm": 0.0553814135491848,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 1.8384971618652344,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26480836236934,
+ "grad_norm": 0.053538259118795395,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 1.8536994457244873,
+ "step": 2467
+ },
+ {
+ "epoch": 34.278745644599304,
+ "grad_norm": 0.05509314313530922,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 1.8386962413787842,
+ "step": 2468
+ },
+ {
+ "epoch": 34.292682926829265,
+ "grad_norm": 0.054552506655454636,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 1.8502833843231201,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30662020905923,
+ "grad_norm": 0.054326511919498444,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 1.8350508213043213,
+ "step": 2470
+ },
+ {
+ "epoch": 34.3205574912892,
+ "grad_norm": 0.05458610877394676,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 1.8488143682479858,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33449477351916,
+ "grad_norm": 0.05537666752934456,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 1.8561567068099976,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34843205574913,
+ "grad_norm": 0.056333623826503754,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 1.8610734939575195,
+ "step": 2473
+ },
+ {
+ "epoch": 34.362369337979096,
+ "grad_norm": 0.053708501160144806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 1.8342688083648682,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37630662020906,
+ "grad_norm": 0.05496754124760628,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 1.8592373132705688,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390243902439025,
+ "grad_norm": 0.05445944517850876,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 1.8551034927368164,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40418118466899,
+ "grad_norm": 0.055870212614536285,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 1.8510769605636597,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41811846689895,
+ "grad_norm": 0.05536561831831932,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 1.8463027477264404,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43205574912892,
+ "grad_norm": 0.054308436810970306,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 1.8470419645309448,
+ "step": 2479
+ },
+ {
+ "epoch": 34.44599303135889,
+ "grad_norm": 0.056683722883462906,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 1.8605597019195557,
+ "step": 2480
+ },
+ {
+ "epoch": 34.45993031358885,
+ "grad_norm": 0.05601316690444946,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 1.847717046737671,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47386759581882,
+ "grad_norm": 0.05449012294411659,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 1.8464031219482422,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48780487804878,
+ "grad_norm": 0.054641272872686386,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 1.832083821296692,
+ "step": 2483
+ },
+ {
+ "epoch": 34.501742160278745,
+ "grad_norm": 0.05610528588294983,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 1.8700275421142578,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51567944250871,
+ "grad_norm": 0.0551300011575222,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 1.8514257669448853,
+ "step": 2485
+ },
+ {
+ "epoch": 34.52961672473867,
+ "grad_norm": 0.056593794375658035,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 1.8480379581451416,
+ "step": 2486
+ },
+ {
+ "epoch": 34.54355400696864,
+ "grad_norm": 0.055673424154520035,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 1.838927984237671,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55749128919861,
+ "grad_norm": 0.05493558943271637,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 1.8383840322494507,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57142857142857,
+ "grad_norm": 0.05680295452475548,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 1.8514118194580078,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58536585365854,
+ "grad_norm": 0.057038743048906326,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 1.8620257377624512,
+ "step": 2490
+ },
+ {
+ "epoch": 34.599303135888505,
+ "grad_norm": 0.05735227093100548,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 1.8413968086242676,
+ "step": 2491
+ },
+ {
+ "epoch": 34.613240418118465,
+ "grad_norm": 0.058245740830898285,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 1.8599750995635986,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62717770034843,
+ "grad_norm": 0.056244585663080215,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 1.8527486324310303,
+ "step": 2493
+ },
+ {
+ "epoch": 34.641114982578394,
+ "grad_norm": 0.05616258084774017,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 1.84233558177948,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65505226480836,
+ "grad_norm": 0.057303108274936676,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 1.865498661994934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.66898954703833,
+ "grad_norm": 0.05895329639315605,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 1.8521645069122314,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68292682926829,
+ "grad_norm": 0.05679212138056755,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 1.8539199829101562,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69686411149826,
+ "grad_norm": 0.05897561088204384,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 1.8534815311431885,
+ "step": 2498
+ },
+ {
+ "epoch": 34.710801393728225,
+ "grad_norm": 0.05891786143183708,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 1.845531702041626,
+ "step": 2499
+ },
+ {
+ "epoch": 34.724738675958186,
+ "grad_norm": 0.059177134186029434,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 1.8502740859985352,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73867595818815,
+ "grad_norm": 0.05862346291542053,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 1.8657329082489014,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75261324041812,
+ "grad_norm": 0.061109986156225204,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 1.8452383279800415,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76655052264808,
+ "grad_norm": 0.058431629091501236,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 1.8575074672698975,
+ "step": 2503
+ },
+ {
+ "epoch": 34.78048780487805,
+ "grad_norm": 0.05970519408583641,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 1.8476191759109497,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79442508710802,
+ "grad_norm": 0.05823173746466637,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 1.8554785251617432,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80836236933798,
+ "grad_norm": 0.060084082186222076,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 1.8471784591674805,
+ "step": 2506
+ },
+ {
+ "epoch": 34.822299651567945,
+ "grad_norm": 0.0589512400329113,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 1.86192786693573,
+ "step": 2507
+ },
+ {
+ "epoch": 34.836236933797906,
+ "grad_norm": 0.06184513866901398,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 1.8572009801864624,
+ "step": 2508
+ },
+ {
+ "epoch": 34.850174216027874,
+ "grad_norm": 0.05833953991532326,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 1.8561553955078125,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86411149825784,
+ "grad_norm": 0.061172984540462494,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 1.8583234548568726,
+ "step": 2510
+ },
+ {
+ "epoch": 34.8780487804878,
+ "grad_norm": 0.0573224313557148,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 1.8403494358062744,
+ "step": 2511
+ },
+ {
+ "epoch": 34.89198606271777,
+ "grad_norm": 0.06135294586420059,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 1.869114875793457,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90592334494774,
+ "grad_norm": 0.05907885357737541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 1.8581886291503906,
+ "step": 2513
+ },
+ {
+ "epoch": 34.9198606271777,
+ "grad_norm": 0.06035003438591957,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 1.8454012870788574,
+ "step": 2514
+ },
+ {
+ "epoch": 34.933797909407666,
+ "grad_norm": 0.060585640370845795,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 1.8453309535980225,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94773519163763,
+ "grad_norm": 0.06050759181380272,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 1.8467144966125488,
+ "step": 2516
+ },
+ {
+ "epoch": 34.961672473867594,
+ "grad_norm": 0.06118239462375641,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 1.8623638153076172,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97560975609756,
+ "grad_norm": 0.059003446251153946,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 1.8540363311767578,
+ "step": 2518
+ },
+ {
+ "epoch": 34.98954703832753,
+ "grad_norm": 0.06078164279460907,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 1.8434364795684814,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04768386855721474,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 1.3819507360458374,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.642964243888855,
+ "eval_runtime": 40.0846,
+ "eval_samples_per_second": 60.921,
+ "eval_steps_per_second": 0.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01393728222997,
+ "grad_norm": 0.05531901493668556,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 1.8454805612564087,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02787456445993,
+ "grad_norm": 0.05443203076720238,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 1.8266499042510986,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041811846689896,
+ "grad_norm": 0.05491992458701134,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 1.839545488357544,
+ "step": 2523
+ },
+ {
+ "epoch": 35.055749128919864,
+ "grad_norm": 0.05408782139420509,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 1.836404800415039,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069686411149824,
+ "grad_norm": 0.0541234128177166,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 1.8264034986495972,
+ "step": 2525
+ },
+ {
+ "epoch": 35.08362369337979,
+ "grad_norm": 0.05504459887742996,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 1.8485287427902222,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09756097560975,
+ "grad_norm": 0.05482470244169235,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 1.8336119651794434,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11149825783972,
+ "grad_norm": 0.05490747466683388,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 1.8223836421966553,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12543554006969,
+ "grad_norm": 0.05421694368124008,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 1.8309847116470337,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13937282229965,
+ "grad_norm": 0.05432746186852455,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 1.8214728832244873,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153310104529616,
+ "grad_norm": 0.05329402908682823,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 1.8356881141662598,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167247386759584,
+ "grad_norm": 0.05515717342495918,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 1.8358293771743774,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181184668989545,
+ "grad_norm": 0.05425635725259781,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 1.8230032920837402,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19512195121951,
+ "grad_norm": 0.054918598383665085,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 1.8427648544311523,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20905923344948,
+ "grad_norm": 0.05530979484319687,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 1.811009407043457,
+ "step": 2535
+ },
+ {
+ "epoch": 35.22299651567944,
+ "grad_norm": 0.05482922121882439,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 1.8356122970581055,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23693379790941,
+ "grad_norm": 0.05470896139740944,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 1.8432835340499878,
+ "step": 2537
+ },
+ {
+ "epoch": 35.250871080139376,
+ "grad_norm": 0.05382648482918739,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 1.8248438835144043,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26480836236934,
+ "grad_norm": 0.05452217534184456,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 1.8385814428329468,
+ "step": 2539
+ },
+ {
+ "epoch": 35.278745644599304,
+ "grad_norm": 0.05492600053548813,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 1.8321887254714966,
+ "step": 2540
+ },
+ {
+ "epoch": 35.292682926829265,
+ "grad_norm": 0.054316446185112,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 1.8411835432052612,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30662020905923,
+ "grad_norm": 0.05490174889564514,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 1.8350071907043457,
+ "step": 2542
+ },
+ {
+ "epoch": 35.3205574912892,
+ "grad_norm": 0.055120889097452164,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 1.8385683298110962,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33449477351916,
+ "grad_norm": 0.054340824484825134,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 1.856696605682373,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34843205574913,
+ "grad_norm": 0.056243546307086945,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 1.8386306762695312,
+ "step": 2545
+ },
+ {
+ "epoch": 35.362369337979096,
+ "grad_norm": 0.05486825481057167,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 1.8275195360183716,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37630662020906,
+ "grad_norm": 0.05750235170125961,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 1.8266764879226685,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390243902439025,
+ "grad_norm": 0.05539723485708237,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 1.8288836479187012,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40418118466899,
+ "grad_norm": 0.056604526937007904,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 1.8384013175964355,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41811846689895,
+ "grad_norm": 0.05410664901137352,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 1.8481371402740479,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43205574912892,
+ "grad_norm": 0.0576784573495388,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 1.8333299160003662,
+ "step": 2551
+ },
+ {
+ "epoch": 35.44599303135889,
+ "grad_norm": 0.05358795076608658,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 1.8201930522918701,
+ "step": 2552
+ },
+ {
+ "epoch": 35.45993031358885,
+ "grad_norm": 0.05687721073627472,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 1.838738203048706,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47386759581882,
+ "grad_norm": 0.05585115775465965,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 1.8358794450759888,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48780487804878,
+ "grad_norm": 0.057034578174352646,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 1.8355228900909424,
+ "step": 2555
+ },
+ {
+ "epoch": 35.501742160278745,
+ "grad_norm": 0.05627685412764549,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 1.8351898193359375,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51567944250871,
+ "grad_norm": 0.05724332481622696,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 1.833801031112671,
+ "step": 2557
+ },
+ {
+ "epoch": 35.52961672473867,
+ "grad_norm": 0.05503929778933525,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 1.8311514854431152,
+ "step": 2558
+ },
+ {
+ "epoch": 35.54355400696864,
+ "grad_norm": 0.057043809443712234,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 1.8318891525268555,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55749128919861,
+ "grad_norm": 0.05669008940458298,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 1.8333967924118042,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57142857142857,
+ "grad_norm": 0.05707597732543945,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 1.8416266441345215,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58536585365854,
+ "grad_norm": 0.055258579552173615,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 1.8428752422332764,
+ "step": 2562
+ },
+ {
+ "epoch": 35.599303135888505,
+ "grad_norm": 0.05575427785515785,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 1.8269420862197876,
+ "step": 2563
+ },
+ {
+ "epoch": 35.613240418118465,
+ "grad_norm": 0.05566711723804474,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 1.8391327857971191,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62717770034843,
+ "grad_norm": 0.057135093957185745,
+ "learning_rate": 0.000125422220031917,
+ "loss": 1.8499560356140137,
+ "step": 2565
+ },
+ {
+ "epoch": 35.641114982578394,
+ "grad_norm": 0.05612964928150177,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 1.8274946212768555,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65505226480836,
+ "grad_norm": 0.05745973810553551,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 1.818924903869629,
+ "step": 2567
+ },
+ {
+ "epoch": 35.66898954703833,
+ "grad_norm": 0.05643341690301895,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 1.8365602493286133,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68292682926829,
+ "grad_norm": 0.054355110973119736,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 1.8275706768035889,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69686411149826,
+ "grad_norm": 0.05827079713344574,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 1.8421577215194702,
+ "step": 2570
+ },
+ {
+ "epoch": 35.710801393728225,
+ "grad_norm": 0.055892039090394974,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 1.8322241306304932,
+ "step": 2571
+ },
+ {
+ "epoch": 35.724738675958186,
+ "grad_norm": 0.05828035622835159,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 1.8444736003875732,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73867595818815,
+ "grad_norm": 0.05669068545103073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 1.8492958545684814,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75261324041812,
+ "grad_norm": 0.05769318342208862,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 1.8362290859222412,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76655052264808,
+ "grad_norm": 0.059012528508901596,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 1.8452755212783813,
+ "step": 2575
+ },
+ {
+ "epoch": 35.78048780487805,
+ "grad_norm": 0.0577283576130867,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 1.8381261825561523,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79442508710802,
+ "grad_norm": 0.05785193666815758,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 1.8344440460205078,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80836236933798,
+ "grad_norm": 0.057778336107730865,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 1.8465049266815186,
+ "step": 2578
+ },
+ {
+ "epoch": 35.822299651567945,
+ "grad_norm": 0.05725527182221413,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 1.8357681035995483,
+ "step": 2579
+ },
+ {
+ "epoch": 35.836236933797906,
+ "grad_norm": 0.05706310644745827,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 1.8430027961730957,
+ "step": 2580
+ },
+ {
+ "epoch": 35.850174216027874,
+ "grad_norm": 0.05754264071583748,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 1.8457331657409668,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86411149825784,
+ "grad_norm": 0.05796574801206589,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 1.8521018028259277,
+ "step": 2582
+ },
+ {
+ "epoch": 35.8780487804878,
+ "grad_norm": 0.05767996609210968,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 1.837098240852356,
+ "step": 2583
+ },
+ {
+ "epoch": 35.89198606271777,
+ "grad_norm": 0.05741669982671738,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 1.8438975811004639,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90592334494774,
+ "grad_norm": 0.05797840282320976,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 1.8390898704528809,
+ "step": 2585
+ },
+ {
+ "epoch": 35.9198606271777,
+ "grad_norm": 0.057005491107702255,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 1.853393316268921,
+ "step": 2586
+ },
+ {
+ "epoch": 35.933797909407666,
+ "grad_norm": 0.058777451515197754,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 1.8450913429260254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94773519163763,
+ "grad_norm": 0.05810392275452614,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 1.8371872901916504,
+ "step": 2588
+ },
+ {
+ "epoch": 35.961672473867594,
+ "grad_norm": 0.05875588208436966,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 1.8611619472503662,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97560975609756,
+ "grad_norm": 0.05872174724936485,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 1.8622443675994873,
+ "step": 2590
+ },
+ {
+ "epoch": 35.98954703832753,
+ "grad_norm": 0.059304554015398026,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 1.8395988941192627,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04685443267226219,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 1.3844654560089111,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.644555926322937,
+ "eval_runtime": 40.5311,
+ "eval_samples_per_second": 60.25,
+ "eval_steps_per_second": 0.493,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01393728222997,
+ "grad_norm": 0.05542989447712898,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 1.816087007522583,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02787456445993,
+ "grad_norm": 0.05269062519073486,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 1.8240373134613037,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041811846689896,
+ "grad_norm": 0.05406419187784195,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 1.8276968002319336,
+ "step": 2595
+ },
+ {
+ "epoch": 36.055749128919864,
+ "grad_norm": 0.052498709410429,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 1.8275878429412842,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069686411149824,
+ "grad_norm": 0.05208680033683777,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 1.8087416887283325,
+ "step": 2597
+ },
+ {
+ "epoch": 36.08362369337979,
+ "grad_norm": 0.05346959829330444,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 1.8090479373931885,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09756097560975,
+ "grad_norm": 0.052402183413505554,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 1.8265495300292969,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11149825783972,
+ "grad_norm": 0.051748018711805344,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 1.8173667192459106,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12543554006969,
+ "grad_norm": 0.05207061767578125,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 1.821485996246338,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13937282229965,
+ "grad_norm": 0.05230815336108208,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 1.804908037185669,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153310104529616,
+ "grad_norm": 0.052593570202589035,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 1.8154864311218262,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167247386759584,
+ "grad_norm": 0.05259322747588158,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 1.8238556385040283,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181184668989545,
+ "grad_norm": 0.052652861922979355,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 1.8147733211517334,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19512195121951,
+ "grad_norm": 0.0530865415930748,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 1.8086535930633545,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20905923344948,
+ "grad_norm": 0.053423281759023666,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 1.8174736499786377,
+ "step": 2607
+ },
+ {
+ "epoch": 36.22299651567944,
+ "grad_norm": 0.0524580255150795,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 1.827101230621338,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23693379790941,
+ "grad_norm": 0.05376284569501877,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 1.8410698175430298,
+ "step": 2609
+ },
+ {
+ "epoch": 36.250871080139376,
+ "grad_norm": 0.05277293920516968,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 1.8262073993682861,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26480836236934,
+ "grad_norm": 0.05415027216076851,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 1.8311805725097656,
+ "step": 2611
+ },
+ {
+ "epoch": 36.278745644599304,
+ "grad_norm": 0.05142449960112572,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 1.816835880279541,
+ "step": 2612
+ },
+ {
+ "epoch": 36.292682926829265,
+ "grad_norm": 0.053121037781238556,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 1.818075180053711,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30662020905923,
+ "grad_norm": 0.051999714225530624,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 1.8257043361663818,
+ "step": 2614
+ },
+ {
+ "epoch": 36.3205574912892,
+ "grad_norm": 0.05257652327418327,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 1.814011573791504,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33449477351916,
+ "grad_norm": 0.0541016161441803,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 1.826779842376709,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34843205574913,
+ "grad_norm": 0.05318204313516617,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 1.8242204189300537,
+ "step": 2617
+ },
+ {
+ "epoch": 36.362369337979096,
+ "grad_norm": 0.05453426018357277,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 1.822621464729309,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37630662020906,
+ "grad_norm": 0.05376817658543587,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 1.8253248929977417,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390243902439025,
+ "grad_norm": 0.05483992025256157,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 1.8317890167236328,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40418118466899,
+ "grad_norm": 0.05439801886677742,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 1.835724949836731,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41811846689895,
+ "grad_norm": 0.055256184190511703,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 1.823319911956787,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43205574912892,
+ "grad_norm": 0.05287355184555054,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 1.8103458881378174,
+ "step": 2623
+ },
+ {
+ "epoch": 36.44599303135889,
+ "grad_norm": 0.05546991154551506,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 1.8309146165847778,
+ "step": 2624
+ },
+ {
+ "epoch": 36.45993031358885,
+ "grad_norm": 0.05322236195206642,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 1.821721076965332,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47386759581882,
+ "grad_norm": 0.054857514798641205,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 1.8251492977142334,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48780487804878,
+ "grad_norm": 0.053303562104701996,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 1.8257558345794678,
+ "step": 2627
+ },
+ {
+ "epoch": 36.501742160278745,
+ "grad_norm": 0.05594215914607048,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 1.835466742515564,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51567944250871,
+ "grad_norm": 0.05624304339289665,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 1.8186707496643066,
+ "step": 2629
+ },
+ {
+ "epoch": 36.52961672473867,
+ "grad_norm": 0.055144522339105606,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 1.8301169872283936,
+ "step": 2630
+ },
+ {
+ "epoch": 36.54355400696864,
+ "grad_norm": 0.0535074919462204,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 1.8138389587402344,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55749128919861,
+ "grad_norm": 0.05484950542449951,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 1.8223397731781006,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57142857142857,
+ "grad_norm": 0.05473049357533455,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 1.8218905925750732,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58536585365854,
+ "grad_norm": 0.05588974803686142,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 1.825666904449463,
+ "step": 2634
+ },
+ {
+ "epoch": 36.599303135888505,
+ "grad_norm": 0.05313410982489586,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 1.8338199853897095,
+ "step": 2635
+ },
+ {
+ "epoch": 36.613240418118465,
+ "grad_norm": 0.056002989411354065,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 1.8219728469848633,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62717770034843,
+ "grad_norm": 0.05519993603229523,
+ "learning_rate": 0.000110418175419276,
+ "loss": 1.826596736907959,
+ "step": 2637
+ },
+ {
+ "epoch": 36.641114982578394,
+ "grad_norm": 0.055067840963602066,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 1.8218865394592285,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65505226480836,
+ "grad_norm": 0.0548541434109211,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 1.8283402919769287,
+ "step": 2639
+ },
+ {
+ "epoch": 36.66898954703833,
+ "grad_norm": 0.05756586790084839,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 1.8259823322296143,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68292682926829,
+ "grad_norm": 0.05620390921831131,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 1.8149542808532715,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69686411149826,
+ "grad_norm": 0.05768077075481415,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 1.8275048732757568,
+ "step": 2642
+ },
+ {
+ "epoch": 36.710801393728225,
+ "grad_norm": 0.05531959980726242,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 1.8295093774795532,
+ "step": 2643
+ },
+ {
+ "epoch": 36.724738675958186,
+ "grad_norm": 0.057147517800331116,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 1.8217309713363647,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73867595818815,
+ "grad_norm": 0.05488502234220505,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 1.8418645858764648,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75261324041812,
+ "grad_norm": 0.05747298523783684,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 1.825081467628479,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76655052264808,
+ "grad_norm": 0.05621277540922165,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 1.8208179473876953,
+ "step": 2647
+ },
+ {
+ "epoch": 36.78048780487805,
+ "grad_norm": 0.05534190684556961,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 1.8269200325012207,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79442508710802,
+ "grad_norm": 0.0570702850818634,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 1.8310284614562988,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80836236933798,
+ "grad_norm": 0.055876970291137695,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 1.831892967224121,
+ "step": 2650
+ },
+ {
+ "epoch": 36.822299651567945,
+ "grad_norm": 0.05680249258875847,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 1.8280737400054932,
+ "step": 2651
+ },
+ {
+ "epoch": 36.836236933797906,
+ "grad_norm": 0.056308578699827194,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 1.844076156616211,
+ "step": 2652
+ },
+ {
+ "epoch": 36.850174216027874,
+ "grad_norm": 0.05752384290099144,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 1.8403749465942383,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86411149825784,
+ "grad_norm": 0.058181654661893845,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 1.8354277610778809,
+ "step": 2654
+ },
+ {
+ "epoch": 36.8780487804878,
+ "grad_norm": 0.05747664347290993,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 1.8315486907958984,
+ "step": 2655
+ },
+ {
+ "epoch": 36.89198606271777,
+ "grad_norm": 0.057498663663864136,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 1.8256535530090332,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90592334494774,
+ "grad_norm": 0.0565464086830616,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 1.8265036344528198,
+ "step": 2657
+ },
+ {
+ "epoch": 36.9198606271777,
+ "grad_norm": 0.057560283690690994,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 1.822776198387146,
+ "step": 2658
+ },
+ {
+ "epoch": 36.933797909407666,
+ "grad_norm": 0.05598663538694382,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 1.8307170867919922,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94773519163763,
+ "grad_norm": 0.0580892413854599,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 1.831838846206665,
+ "step": 2660
+ },
+ {
+ "epoch": 36.961672473867594,
+ "grad_norm": 0.056303758174180984,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 1.837568998336792,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97560975609756,
+ "grad_norm": 0.05781455338001251,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 1.8347523212432861,
+ "step": 2662
+ },
+ {
+ "epoch": 36.98954703832753,
+ "grad_norm": 0.05536883324384689,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 1.8196232318878174,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.046378862112760544,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 1.3591325283050537,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6459980607032776,
+ "eval_runtime": 40.631,
+ "eval_samples_per_second": 60.102,
+ "eval_steps_per_second": 0.492,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01393728222997,
+ "grad_norm": 0.050504088401794434,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 1.8197872638702393,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02787456445993,
+ "grad_norm": 0.05133911967277527,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 1.8092283010482788,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041811846689896,
+ "grad_norm": 0.051401812583208084,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 1.7972350120544434,
+ "step": 2667
+ },
+ {
+ "epoch": 37.055749128919864,
+ "grad_norm": 0.050701867789030075,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 1.8077106475830078,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069686411149824,
+ "grad_norm": 0.05219056457281113,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 1.7995078563690186,
+ "step": 2669
+ },
+ {
+ "epoch": 37.08362369337979,
+ "grad_norm": 0.05003709718585014,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 1.8166279792785645,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09756097560975,
+ "grad_norm": 0.051945704966783524,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 1.8085274696350098,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11149825783972,
+ "grad_norm": 0.05073661729693413,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 1.794724941253662,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12543554006969,
+ "grad_norm": 0.051837258040905,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 1.8101842403411865,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13937282229965,
+ "grad_norm": 0.051624711602926254,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 1.8128712177276611,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153310104529616,
+ "grad_norm": 0.05337493121623993,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 1.8105556964874268,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167247386759584,
+ "grad_norm": 0.05040125921368599,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 1.801112413406372,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181184668989545,
+ "grad_norm": 0.05222955718636513,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 1.8220586776733398,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19512195121951,
+ "grad_norm": 0.052356619387865067,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 1.8208140134811401,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20905923344948,
+ "grad_norm": 0.05221114307641983,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 1.8082315921783447,
+ "step": 2679
+ },
+ {
+ "epoch": 37.22299651567944,
+ "grad_norm": 0.05170206353068352,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 1.8159395456314087,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23693379790941,
+ "grad_norm": 0.05091318488121033,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 1.8138477802276611,
+ "step": 2681
+ },
+ {
+ "epoch": 37.250871080139376,
+ "grad_norm": 0.05228526517748833,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 1.8149440288543701,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26480836236934,
+ "grad_norm": 0.053375136107206345,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 1.8169589042663574,
+ "step": 2683
+ },
+ {
+ "epoch": 37.278745644599304,
+ "grad_norm": 0.05215807631611824,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 1.8025376796722412,
+ "step": 2684
+ },
+ {
+ "epoch": 37.292682926829265,
+ "grad_norm": 0.05166545882821083,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 1.815457820892334,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30662020905923,
+ "grad_norm": 0.05314334109425545,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 1.8092843294143677,
+ "step": 2686
+ },
+ {
+ "epoch": 37.3205574912892,
+ "grad_norm": 0.0525335855782032,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 1.8068681955337524,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33449477351916,
+ "grad_norm": 0.05154058337211609,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 1.8220410346984863,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34843205574913,
+ "grad_norm": 0.05400070548057556,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 1.8106049299240112,
+ "step": 2689
+ },
+ {
+ "epoch": 37.362369337979096,
+ "grad_norm": 0.051827941089868546,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 1.8139021396636963,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37630662020906,
+ "grad_norm": 0.053685542196035385,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 1.8051722049713135,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390243902439025,
+ "grad_norm": 0.05291858687996864,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 1.806849479675293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40418118466899,
+ "grad_norm": 0.05310729518532753,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 1.7940170764923096,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41811846689895,
+ "grad_norm": 0.05288896709680557,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 1.824284553527832,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43205574912892,
+ "grad_norm": 0.052886299788951874,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 1.8206238746643066,
+ "step": 2695
+ },
+ {
+ "epoch": 37.44599303135889,
+ "grad_norm": 0.05356050282716751,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 1.8159115314483643,
+ "step": 2696
+ },
+ {
+ "epoch": 37.45993031358885,
+ "grad_norm": 0.053028833121061325,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 1.8148610591888428,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47386759581882,
+ "grad_norm": 0.05355606600642204,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 1.8167890310287476,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48780487804878,
+ "grad_norm": 0.053764909505844116,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 1.8095712661743164,
+ "step": 2699
+ },
+ {
+ "epoch": 37.501742160278745,
+ "grad_norm": 0.05224880203604698,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 1.8151801824569702,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51567944250871,
+ "grad_norm": 0.05442487820982933,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 1.8090888261795044,
+ "step": 2701
+ },
+ {
+ "epoch": 37.52961672473867,
+ "grad_norm": 0.05287255719304085,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 1.8190618753433228,
+ "step": 2702
+ },
+ {
+ "epoch": 37.54355400696864,
+ "grad_norm": 0.05432182550430298,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 1.8103792667388916,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55749128919861,
+ "grad_norm": 0.05345293879508972,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 1.8107848167419434,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57142857142857,
+ "grad_norm": 0.05567241832613945,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 1.822261095046997,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58536585365854,
+ "grad_norm": 0.0534495934844017,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 1.8195551633834839,
+ "step": 2706
+ },
+ {
+ "epoch": 37.599303135888505,
+ "grad_norm": 0.05432790517807007,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 1.8119161128997803,
+ "step": 2707
+ },
+ {
+ "epoch": 37.613240418118465,
+ "grad_norm": 0.05465250089764595,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 1.8269178867340088,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62717770034843,
+ "grad_norm": 0.05540885776281357,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 1.8170043230056763,
+ "step": 2709
+ },
+ {
+ "epoch": 37.641114982578394,
+ "grad_norm": 0.053956300020217896,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 1.8141452074050903,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65505226480836,
+ "grad_norm": 0.0530884824693203,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 1.8228175640106201,
+ "step": 2711
+ },
+ {
+ "epoch": 37.66898954703833,
+ "grad_norm": 0.05414469167590141,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 1.8148912191390991,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68292682926829,
+ "grad_norm": 0.054135460406541824,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 1.8189573287963867,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69686411149826,
+ "grad_norm": 0.05427670478820801,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 1.8352646827697754,
+ "step": 2714
+ },
+ {
+ "epoch": 37.710801393728225,
+ "grad_norm": 0.054425179958343506,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 1.8101688623428345,
+ "step": 2715
+ },
+ {
+ "epoch": 37.724738675958186,
+ "grad_norm": 0.05361952632665634,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 1.822049856185913,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73867595818815,
+ "grad_norm": 0.053644467145204544,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 1.8222166299819946,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75261324041812,
+ "grad_norm": 0.054326146841049194,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 1.8122522830963135,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76655052264808,
+ "grad_norm": 0.05346079170703888,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 1.8157234191894531,
+ "step": 2719
+ },
+ {
+ "epoch": 37.78048780487805,
+ "grad_norm": 0.05502757802605629,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 1.8146154880523682,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79442508710802,
+ "grad_norm": 0.05353285372257233,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 1.81754732131958,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80836236933798,
+ "grad_norm": 0.05462120845913887,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 1.8201050758361816,
+ "step": 2722
+ },
+ {
+ "epoch": 37.822299651567945,
+ "grad_norm": 0.05272475257515907,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 1.8173773288726807,
+ "step": 2723
+ },
+ {
+ "epoch": 37.836236933797906,
+ "grad_norm": 0.056003544479608536,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 1.8211115598678589,
+ "step": 2724
+ },
+ {
+ "epoch": 37.850174216027874,
+ "grad_norm": 0.05486152321100235,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 1.8221628665924072,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86411149825784,
+ "grad_norm": 0.05549563467502594,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 1.8200361728668213,
+ "step": 2726
+ },
+ {
+ "epoch": 37.8780487804878,
+ "grad_norm": 0.05407272279262543,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 1.814354658126831,
+ "step": 2727
+ },
+ {
+ "epoch": 37.89198606271777,
+ "grad_norm": 0.05526091903448105,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 1.8025412559509277,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90592334494774,
+ "grad_norm": 0.05579185113310814,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 1.8137646913528442,
+ "step": 2729
+ },
+ {
+ "epoch": 37.9198606271777,
+ "grad_norm": 0.054667308926582336,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 1.8100636005401611,
+ "step": 2730
+ },
+ {
+ "epoch": 37.933797909407666,
+ "grad_norm": 0.054524678736925125,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 1.8058973550796509,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94773519163763,
+ "grad_norm": 0.05454818904399872,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 1.8153016567230225,
+ "step": 2732
+ },
+ {
+ "epoch": 37.961672473867594,
+ "grad_norm": 0.05491559952497482,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 1.8142199516296387,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97560975609756,
+ "grad_norm": 0.055562835186719894,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 1.8082599639892578,
+ "step": 2734
+ },
+ {
+ "epoch": 37.98954703832753,
+ "grad_norm": 0.055928416550159454,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 1.8185434341430664,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.044970303773880005,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 1.346644401550293,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6474234461784363,
+ "eval_runtime": 40.6464,
+ "eval_samples_per_second": 60.079,
+ "eval_steps_per_second": 0.492,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01393728222997,
+ "grad_norm": 0.05179993808269501,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 1.7899185419082642,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02787456445993,
+ "grad_norm": 0.04930533468723297,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 1.7904696464538574,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041811846689896,
+ "grad_norm": 0.052794501185417175,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 1.7964375019073486,
+ "step": 2739
+ },
+ {
+ "epoch": 38.055749128919864,
+ "grad_norm": 0.049202270805835724,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 1.790614128112793,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069686411149824,
+ "grad_norm": 0.04903653636574745,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 1.803176999092102,
+ "step": 2741
+ },
+ {
+ "epoch": 38.08362369337979,
+ "grad_norm": 0.04976186528801918,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 1.8021293878555298,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09756097560975,
+ "grad_norm": 0.05226270109415054,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 1.7997218370437622,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11149825783972,
+ "grad_norm": 0.04967877268791199,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 1.8076910972595215,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12543554006969,
+ "grad_norm": 0.05012716352939606,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 1.7991023063659668,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13937282229965,
+ "grad_norm": 0.05108457803726196,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 1.8080018758773804,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153310104529616,
+ "grad_norm": 0.05005490034818649,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 1.7945563793182373,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167247386759584,
+ "grad_norm": 0.050113365054130554,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 1.802520990371704,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181184668989545,
+ "grad_norm": 0.05096133425831795,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 1.7994227409362793,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19512195121951,
+ "grad_norm": 0.05119607597589493,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 1.8115901947021484,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20905923344948,
+ "grad_norm": 0.051627811044454575,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 1.800724744796753,
+ "step": 2751
+ },
+ {
+ "epoch": 38.22299651567944,
+ "grad_norm": 0.05053049698472023,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 1.7860815525054932,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23693379790941,
+ "grad_norm": 0.05122068524360657,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 1.7976000308990479,
+ "step": 2753
+ },
+ {
+ "epoch": 38.250871080139376,
+ "grad_norm": 0.05197495222091675,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 1.811307430267334,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26480836236934,
+ "grad_norm": 0.051113326102495193,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 1.7974506616592407,
+ "step": 2755
+ },
+ {
+ "epoch": 38.278745644599304,
+ "grad_norm": 0.05044601857662201,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 1.8004801273345947,
+ "step": 2756
+ },
+ {
+ "epoch": 38.292682926829265,
+ "grad_norm": 0.05080671235918999,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 1.7960381507873535,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30662020905923,
+ "grad_norm": 0.05112270265817642,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 1.8070579767227173,
+ "step": 2758
+ },
+ {
+ "epoch": 38.3205574912892,
+ "grad_norm": 0.050059717148542404,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 1.7989699840545654,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33449477351916,
+ "grad_norm": 0.05151006579399109,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 1.793050765991211,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34843205574913,
+ "grad_norm": 0.05045556277036667,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 1.7945470809936523,
+ "step": 2761
+ },
+ {
+ "epoch": 38.362369337979096,
+ "grad_norm": 0.05152078717947006,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 1.8155558109283447,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37630662020906,
+ "grad_norm": 0.049696408212184906,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 1.7929422855377197,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390243902439025,
+ "grad_norm": 0.050947509706020355,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 1.8029417991638184,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40418118466899,
+ "grad_norm": 0.04988504573702812,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 1.807220697402954,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41811846689895,
+ "grad_norm": 0.05031213164329529,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 1.820298433303833,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43205574912892,
+ "grad_norm": 0.05067938566207886,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 1.8025063276290894,
+ "step": 2767
+ },
+ {
+ "epoch": 38.44599303135889,
+ "grad_norm": 0.050599709153175354,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 1.7912819385528564,
+ "step": 2768
+ },
+ {
+ "epoch": 38.45993031358885,
+ "grad_norm": 0.05000261962413788,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 1.8116015195846558,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47386759581882,
+ "grad_norm": 0.05164877697825432,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 1.7894442081451416,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48780487804878,
+ "grad_norm": 0.05037420988082886,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 1.8002345561981201,
+ "step": 2771
+ },
+ {
+ "epoch": 38.501742160278745,
+ "grad_norm": 0.05061398074030876,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 1.8103997707366943,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51567944250871,
+ "grad_norm": 0.05141839757561684,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 1.7920000553131104,
+ "step": 2773
+ },
+ {
+ "epoch": 38.52961672473867,
+ "grad_norm": 0.050547219812870026,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 1.8011581897735596,
+ "step": 2774
+ },
+ {
+ "epoch": 38.54355400696864,
+ "grad_norm": 0.052665967494249344,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 1.8171170949935913,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55749128919861,
+ "grad_norm": 0.05148577317595482,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 1.7976020574569702,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57142857142857,
+ "grad_norm": 0.05378367006778717,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 1.8061752319335938,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58536585365854,
+ "grad_norm": 0.05071459710597992,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 1.8084588050842285,
+ "step": 2778
+ },
+ {
+ "epoch": 38.599303135888505,
+ "grad_norm": 0.05229831859469414,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 1.8100000619888306,
+ "step": 2779
+ },
+ {
+ "epoch": 38.613240418118465,
+ "grad_norm": 0.052434664219617844,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 1.8119068145751953,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62717770034843,
+ "grad_norm": 0.05269220098853111,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 1.799889087677002,
+ "step": 2781
+ },
+ {
+ "epoch": 38.641114982578394,
+ "grad_norm": 0.05271878093481064,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 1.792294979095459,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65505226480836,
+ "grad_norm": 0.05291058123111725,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 1.7999838590621948,
+ "step": 2783
+ },
+ {
+ "epoch": 38.66898954703833,
+ "grad_norm": 0.05346683785319328,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 1.801847219467163,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68292682926829,
+ "grad_norm": 0.053118057548999786,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 1.7951825857162476,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69686411149826,
+ "grad_norm": 0.0527576319873333,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 1.8107949495315552,
+ "step": 2786
+ },
+ {
+ "epoch": 38.710801393728225,
+ "grad_norm": 0.05188573896884918,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 1.8066956996917725,
+ "step": 2787
+ },
+ {
+ "epoch": 38.724738675958186,
+ "grad_norm": 0.05341636762022972,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 1.8036713600158691,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73867595818815,
+ "grad_norm": 0.051814883947372437,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 1.8106606006622314,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75261324041812,
+ "grad_norm": 0.05374172329902649,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 1.7904407978057861,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76655052264808,
+ "grad_norm": 0.052160196006298065,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 1.797128677368164,
+ "step": 2791
+ },
+ {
+ "epoch": 38.78048780487805,
+ "grad_norm": 0.052794717252254486,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 1.8085696697235107,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79442508710802,
+ "grad_norm": 0.05286112427711487,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 1.8027613162994385,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80836236933798,
+ "grad_norm": 0.052660733461380005,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 1.8135063648223877,
+ "step": 2794
+ },
+ {
+ "epoch": 38.822299651567945,
+ "grad_norm": 0.052824389189481735,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 1.7999730110168457,
+ "step": 2795
+ },
+ {
+ "epoch": 38.836236933797906,
+ "grad_norm": 0.05390750244259834,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 1.7879369258880615,
+ "step": 2796
+ },
+ {
+ "epoch": 38.850174216027874,
+ "grad_norm": 0.053744010627269745,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 1.811363935470581,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86411149825784,
+ "grad_norm": 0.05401702970266342,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 1.8049407005310059,
+ "step": 2798
+ },
+ {
+ "epoch": 38.8780487804878,
+ "grad_norm": 0.05376699939370155,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 1.8111652135849,
+ "step": 2799
+ },
+ {
+ "epoch": 38.89198606271777,
+ "grad_norm": 0.05381734296679497,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 1.8312594890594482,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90592334494774,
+ "grad_norm": 0.05387615039944649,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 1.805098295211792,
+ "step": 2801
+ },
+ {
+ "epoch": 38.9198606271777,
+ "grad_norm": 0.05285770446062088,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 1.8160698413848877,
+ "step": 2802
+ },
+ {
+ "epoch": 38.933797909407666,
+ "grad_norm": 0.05545372888445854,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 1.8070565462112427,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94773519163763,
+ "grad_norm": 0.05404449254274368,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 1.8055229187011719,
+ "step": 2804
+ },
+ {
+ "epoch": 38.961672473867594,
+ "grad_norm": 0.05500908941030502,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 1.798984169960022,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97560975609756,
+ "grad_norm": 0.05408451333642006,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 1.8082829713821411,
+ "step": 2806
+ },
+ {
+ "epoch": 38.98954703832753,
+ "grad_norm": 0.0534386970102787,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 1.8156507015228271,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.044039830565452576,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 1.3510627746582031,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6490727663040161,
+ "eval_runtime": 41.0373,
+ "eval_samples_per_second": 59.507,
+ "eval_steps_per_second": 0.487,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01393728222997,
+ "grad_norm": 0.0485052689909935,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 1.787034511566162,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02787456445993,
+ "grad_norm": 0.046773314476013184,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 1.7824962139129639,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041811846689896,
+ "grad_norm": 0.04776127636432648,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 1.784999132156372,
+ "step": 2811
+ },
+ {
+ "epoch": 39.055749128919864,
+ "grad_norm": 0.04702889546751976,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 1.7887163162231445,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069686411149824,
+ "grad_norm": 0.0472981221973896,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 1.7778058052062988,
+ "step": 2813
+ },
+ {
+ "epoch": 39.08362369337979,
+ "grad_norm": 0.04829077050089836,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 1.8164949417114258,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09756097560975,
+ "grad_norm": 0.04883906990289688,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 1.787117838859558,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11149825783972,
+ "grad_norm": 0.0464581660926342,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 1.7883102893829346,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12543554006969,
+ "grad_norm": 0.047438278794288635,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 1.7923827171325684,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13937282229965,
+ "grad_norm": 0.04703788831830025,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 1.78762686252594,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153310104529616,
+ "grad_norm": 0.049375683069229126,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 1.7970117330551147,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167247386759584,
+ "grad_norm": 0.04796822741627693,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 1.778883457183838,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181184668989545,
+ "grad_norm": 0.04719251021742821,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 1.7749544382095337,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19512195121951,
+ "grad_norm": 0.04893529415130615,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 1.787841796875,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20905923344948,
+ "grad_norm": 0.04869271442294121,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 1.8007760047912598,
+ "step": 2823
+ },
+ {
+ "epoch": 39.22299651567944,
+ "grad_norm": 0.04950389638543129,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 1.7905025482177734,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23693379790941,
+ "grad_norm": 0.0489855520427227,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 1.7977821826934814,
+ "step": 2825
+ },
+ {
+ "epoch": 39.250871080139376,
+ "grad_norm": 0.04881766065955162,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 1.7955210208892822,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26480836236934,
+ "grad_norm": 0.04834857955574989,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 1.7897529602050781,
+ "step": 2827
+ },
+ {
+ "epoch": 39.278745644599304,
+ "grad_norm": 0.04898613691329956,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 1.7888939380645752,
+ "step": 2828
+ },
+ {
+ "epoch": 39.292682926829265,
+ "grad_norm": 0.04936131089925766,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 1.8096022605895996,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30662020905923,
+ "grad_norm": 0.04834993928670883,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 1.800262689590454,
+ "step": 2830
+ },
+ {
+ "epoch": 39.3205574912892,
+ "grad_norm": 0.049713220447301865,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 1.7950725555419922,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33449477351916,
+ "grad_norm": 0.049898866564035416,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 1.7812716960906982,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34843205574913,
+ "grad_norm": 0.049175627529621124,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 1.7746937274932861,
+ "step": 2833
+ },
+ {
+ "epoch": 39.362369337979096,
+ "grad_norm": 0.04872290790081024,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 1.7824972867965698,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37630662020906,
+ "grad_norm": 0.049543045461177826,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 1.793465256690979,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390243902439025,
+ "grad_norm": 0.049746304750442505,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 1.792328119277954,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40418118466899,
+ "grad_norm": 0.04971722885966301,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 1.7932684421539307,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41811846689895,
+ "grad_norm": 0.05128713324666023,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 1.8031436204910278,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43205574912892,
+ "grad_norm": 0.04928988218307495,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 1.7928621768951416,
+ "step": 2839
+ },
+ {
+ "epoch": 39.44599303135889,
+ "grad_norm": 0.05211673676967621,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 1.7749731540679932,
+ "step": 2840
+ },
+ {
+ "epoch": 39.45993031358885,
+ "grad_norm": 0.05039450526237488,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 1.7910631895065308,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47386759581882,
+ "grad_norm": 0.05047272890806198,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 1.8027098178863525,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48780487804878,
+ "grad_norm": 0.05042344331741333,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 1.8134506940841675,
+ "step": 2843
+ },
+ {
+ "epoch": 39.501742160278745,
+ "grad_norm": 0.0506373755633831,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 1.7785459756851196,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51567944250871,
+ "grad_norm": 0.04902270808815956,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 1.7985787391662598,
+ "step": 2845
+ },
+ {
+ "epoch": 39.52961672473867,
+ "grad_norm": 0.05162668600678444,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 1.7813456058502197,
+ "step": 2846
+ },
+ {
+ "epoch": 39.54355400696864,
+ "grad_norm": 0.05005355924367905,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 1.791803240776062,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55749128919861,
+ "grad_norm": 0.050947558134794235,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 1.7949843406677246,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57142857142857,
+ "grad_norm": 0.05130480229854584,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 1.7962243556976318,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58536585365854,
+ "grad_norm": 0.049898259341716766,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 1.7940798997879028,
+ "step": 2850
+ },
+ {
+ "epoch": 39.599303135888505,
+ "grad_norm": 0.050370268523693085,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 1.8001989126205444,
+ "step": 2851
+ },
+ {
+ "epoch": 39.613240418118465,
+ "grad_norm": 0.05121270939707756,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 1.79237961769104,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62717770034843,
+ "grad_norm": 0.05014092102646828,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 1.804434895515442,
+ "step": 2853
+ },
+ {
+ "epoch": 39.641114982578394,
+ "grad_norm": 0.050814609974622726,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 1.7997581958770752,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65505226480836,
+ "grad_norm": 0.05023293197154999,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 1.8026623725891113,
+ "step": 2855
+ },
+ {
+ "epoch": 39.66898954703833,
+ "grad_norm": 0.051415517926216125,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 1.8061354160308838,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68292682926829,
+ "grad_norm": 0.04993635416030884,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 1.7835321426391602,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69686411149826,
+ "grad_norm": 0.05157633125782013,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 1.8007519245147705,
+ "step": 2858
+ },
+ {
+ "epoch": 39.710801393728225,
+ "grad_norm": 0.050889477133750916,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 1.7806460857391357,
+ "step": 2859
+ },
+ {
+ "epoch": 39.724738675958186,
+ "grad_norm": 0.050646744668483734,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 1.780684471130371,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73867595818815,
+ "grad_norm": 0.05104859918355942,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 1.796810507774353,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75261324041812,
+ "grad_norm": 0.051958777010440826,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 1.8104615211486816,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76655052264808,
+ "grad_norm": 0.05137504264712334,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 1.7988557815551758,
+ "step": 2863
+ },
+ {
+ "epoch": 39.78048780487805,
+ "grad_norm": 0.051518820226192474,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 1.7865259647369385,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79442508710802,
+ "grad_norm": 0.051753368228673935,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 1.8039045333862305,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80836236933798,
+ "grad_norm": 0.0514046885073185,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 1.8040326833724976,
+ "step": 2866
+ },
+ {
+ "epoch": 39.822299651567945,
+ "grad_norm": 0.05259102210402489,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 1.8034707307815552,
+ "step": 2867
+ },
+ {
+ "epoch": 39.836236933797906,
+ "grad_norm": 0.05125397443771362,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 1.7860667705535889,
+ "step": 2868
+ },
+ {
+ "epoch": 39.850174216027874,
+ "grad_norm": 0.05120245739817619,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 1.7920024394989014,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86411149825784,
+ "grad_norm": 0.05269347131252289,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 1.7958977222442627,
+ "step": 2870
+ },
+ {
+ "epoch": 39.8780487804878,
+ "grad_norm": 0.050947729498147964,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 1.7864038944244385,
+ "step": 2871
+ },
+ {
+ "epoch": 39.89198606271777,
+ "grad_norm": 0.052531689405441284,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 1.795427918434143,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90592334494774,
+ "grad_norm": 0.052431896328926086,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 1.7928128242492676,
+ "step": 2873
+ },
+ {
+ "epoch": 39.9198606271777,
+ "grad_norm": 0.053628548979759216,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 1.7986962795257568,
+ "step": 2874
+ },
+ {
+ "epoch": 39.933797909407666,
+ "grad_norm": 0.05186983942985535,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 1.7967369556427002,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94773519163763,
+ "grad_norm": 0.05122850835323334,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 1.7836300134658813,
+ "step": 2876
+ },
+ {
+ "epoch": 39.961672473867594,
+ "grad_norm": 0.05265440791845322,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 1.801450490951538,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97560975609756,
+ "grad_norm": 0.053320810198783875,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 1.792264461517334,
+ "step": 2878
+ },
+ {
+ "epoch": 39.98954703832753,
+ "grad_norm": 0.05239804461598396,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 1.8015801906585693,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04351874440908432,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 1.3430795669555664,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.231062406570967e+18,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-2880/training_args.bin b/runs/mt5-ssl/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-3240/chat_template.jinja b/runs/mt5-ssl/checkpoint-3240/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-3240/config.json b/runs/mt5-ssl/checkpoint-3240/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-3240/generation_config.json b/runs/mt5-ssl/checkpoint-3240/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-3240/model.safetensors b/runs/mt5-ssl/checkpoint-3240/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f9a39040333bd37f1adad73cf4c0f56cc99e1774
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ef51198e7e37f0d478ba80fe6d6fee88ff1d99f8fee82d064baf2e9dfafd7d13
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-3240/optimizer.pt b/runs/mt5-ssl/checkpoint-3240/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ce95c146a4d1bf07ee898693dfec47e00df0f2c8
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1d1254c0fecce6d2e1e8c732590b3380cd2f518a700d06d824172a62506974e8
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-3240/rng_state_0.pth b/runs/mt5-ssl/checkpoint-3240/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..854d237a4ff7656e8561049ef131e53b73d26e13
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:be1f0fdefddf6f30886458ed628f8251f23fec39a294e6649a7ab241784b8460
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-3240/rng_state_1.pth b/runs/mt5-ssl/checkpoint-3240/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bf9f97063d7490c565f537598baa2da46c4b6b26
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:35669f76342834bd66ba665f73a6cfdbe07c4f2c0f41f6e5e6ed2ad3ae67a161
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-3240/scheduler.pt b/runs/mt5-ssl/checkpoint-3240/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..26a875c54fc41f3b55217f78e9787c3a1beec0e7
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ffe46f59a068a92a009d9e85177c3c1b0af06eb72511444aa28197cfb8d5f1be
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-3240/tokenizer.json b/runs/mt5-ssl/checkpoint-3240/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-3240/tokenizer_config.json b/runs/mt5-ssl/checkpoint-3240/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-3240/trainer_state.json b/runs/mt5-ssl/checkpoint-3240/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..7d948879eff62620c279170de3cde93ec88b5b1c
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/trainer_state.json
@@ -0,0 +1,23066 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 45.0,
+ "eval_steps": 500,
+ "global_step": 3240,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6287577152252197,
+ "eval_runtime": 39.8508,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013937282229964,
+ "grad_norm": 0.056210409849882126,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 1.9758458137512207,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027874564459932,
+ "grad_norm": 0.06387963145971298,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 1.9743283987045288,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041811846689896,
+ "grad_norm": 0.059931688010692596,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 1.983224630355835,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05574912891986,
+ "grad_norm": 0.059698957949876785,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 1.9638850688934326,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069686411149824,
+ "grad_norm": 0.057556964457035065,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 1.9888463020324707,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083623693379792,
+ "grad_norm": 0.06250642985105515,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 1.9776880741119385,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097560975609756,
+ "grad_norm": 0.06218443810939789,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 1.9904823303222656,
+ "step": 1807
+ },
+ {
+ "epoch": 25.11149825783972,
+ "grad_norm": 0.06147411838173866,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 1.9881136417388916,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125435540069688,
+ "grad_norm": 0.06080890819430351,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 1.9833264350891113,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139372822299652,
+ "grad_norm": 0.057465583086013794,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 1.9764835834503174,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153310104529616,
+ "grad_norm": 0.061488356441259384,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 1.9807426929473877,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16724738675958,
+ "grad_norm": 0.060737013816833496,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 1.9796113967895508,
+ "step": 1812
+ },
+ {
+ "epoch": 25.181184668989548,
+ "grad_norm": 0.060508858412504196,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 1.9862518310546875,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195121951219512,
+ "grad_norm": 0.06062212213873863,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 1.9855456352233887,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209059233449477,
+ "grad_norm": 0.06143258139491081,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 1.9763805866241455,
+ "step": 1815
+ },
+ {
+ "epoch": 25.222996515679444,
+ "grad_norm": 0.06446489691734314,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 1.9674923419952393,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23693379790941,
+ "grad_norm": 0.06179432198405266,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 1.9792906045913696,
+ "step": 1817
+ },
+ {
+ "epoch": 25.250871080139373,
+ "grad_norm": 0.062219955027103424,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 1.983303189277649,
+ "step": 1818
+ },
+ {
+ "epoch": 25.264808362369337,
+ "grad_norm": 0.060631707310676575,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 1.9859293699264526,
+ "step": 1819
+ },
+ {
+ "epoch": 25.278745644599304,
+ "grad_norm": 0.06329842656850815,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 1.988373041152954,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29268292682927,
+ "grad_norm": 0.06290270388126373,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 1.9817759990692139,
+ "step": 1821
+ },
+ {
+ "epoch": 25.306620209059233,
+ "grad_norm": 0.062234487384557724,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 1.9615445137023926,
+ "step": 1822
+ },
+ {
+ "epoch": 25.320557491289197,
+ "grad_norm": 0.06063779070973396,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 1.9736756086349487,
+ "step": 1823
+ },
+ {
+ "epoch": 25.334494773519165,
+ "grad_norm": 0.06245938315987587,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 1.9932905435562134,
+ "step": 1824
+ },
+ {
+ "epoch": 25.34843205574913,
+ "grad_norm": 0.06380622833967209,
+ "learning_rate": 0.000309930621834911,
+ "loss": 1.9695440530776978,
+ "step": 1825
+ },
+ {
+ "epoch": 25.362369337979093,
+ "grad_norm": 0.0626329705119133,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 1.9820970296859741,
+ "step": 1826
+ },
+ {
+ "epoch": 25.37630662020906,
+ "grad_norm": 0.060844022780656815,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 1.9926050901412964,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390243902439025,
+ "grad_norm": 0.06239933893084526,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 2.001865863800049,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40418118466899,
+ "grad_norm": 0.06491564959287643,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 1.9897377490997314,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418118466898953,
+ "grad_norm": 0.05946580320596695,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 1.973170280456543,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43205574912892,
+ "grad_norm": 0.05930647253990173,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 1.9949123859405518,
+ "step": 1831
+ },
+ {
+ "epoch": 25.445993031358885,
+ "grad_norm": 0.06269685924053192,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 1.9912991523742676,
+ "step": 1832
+ },
+ {
+ "epoch": 25.45993031358885,
+ "grad_norm": 0.0637395903468132,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 1.9869414567947388,
+ "step": 1833
+ },
+ {
+ "epoch": 25.473867595818817,
+ "grad_norm": 0.0628778263926506,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 1.9748289585113525,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48780487804878,
+ "grad_norm": 0.06064455583691597,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 2.008216381072998,
+ "step": 1835
+ },
+ {
+ "epoch": 25.501742160278745,
+ "grad_norm": 0.0655415952205658,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 1.9852811098098755,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51567944250871,
+ "grad_norm": 0.06189798563718796,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 1.983716368675232,
+ "step": 1837
+ },
+ {
+ "epoch": 25.529616724738677,
+ "grad_norm": 0.061454080045223236,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 2.013745069503784,
+ "step": 1838
+ },
+ {
+ "epoch": 25.54355400696864,
+ "grad_norm": 0.06712181121110916,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 1.9856178760528564,
+ "step": 1839
+ },
+ {
+ "epoch": 25.557491289198605,
+ "grad_norm": 0.06364461034536362,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 2.0034379959106445,
+ "step": 1840
+ },
+ {
+ "epoch": 25.571428571428573,
+ "grad_norm": 0.06669855862855911,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 1.9868637323379517,
+ "step": 1841
+ },
+ {
+ "epoch": 25.585365853658537,
+ "grad_norm": 0.06578591465950012,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 1.9998185634613037,
+ "step": 1842
+ },
+ {
+ "epoch": 25.5993031358885,
+ "grad_norm": 0.06305381655693054,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 2.000594139099121,
+ "step": 1843
+ },
+ {
+ "epoch": 25.613240418118465,
+ "grad_norm": 0.061369556933641434,
+ "learning_rate": 0.000304866093757771,
+ "loss": 1.9811584949493408,
+ "step": 1844
+ },
+ {
+ "epoch": 25.627177700348433,
+ "grad_norm": 0.062390975654125214,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 2.0020859241485596,
+ "step": 1845
+ },
+ {
+ "epoch": 25.641114982578397,
+ "grad_norm": 0.06477145850658417,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 2.004000186920166,
+ "step": 1846
+ },
+ {
+ "epoch": 25.65505226480836,
+ "grad_norm": 0.06288585811853409,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 1.9919252395629883,
+ "step": 1847
+ },
+ {
+ "epoch": 25.66898954703833,
+ "grad_norm": 0.05959830805659294,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 1.9860963821411133,
+ "step": 1848
+ },
+ {
+ "epoch": 25.682926829268293,
+ "grad_norm": 0.0596182718873024,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 2.0014610290527344,
+ "step": 1849
+ },
+ {
+ "epoch": 25.696864111498257,
+ "grad_norm": 0.06665527820587158,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 2.0072989463806152,
+ "step": 1850
+ },
+ {
+ "epoch": 25.71080139372822,
+ "grad_norm": 0.0631684735417366,
+ "learning_rate": 0.000303,
+ "loss": 2.0076632499694824,
+ "step": 1851
+ },
+ {
+ "epoch": 25.72473867595819,
+ "grad_norm": 0.06214272975921631,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 2.0143537521362305,
+ "step": 1852
+ },
+ {
+ "epoch": 25.738675958188153,
+ "grad_norm": 0.06422685086727142,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 1.9885808229446411,
+ "step": 1853
+ },
+ {
+ "epoch": 25.752613240418118,
+ "grad_norm": 0.06583613157272339,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 2.0042953491210938,
+ "step": 1854
+ },
+ {
+ "epoch": 25.766550522648085,
+ "grad_norm": 0.06265029311180115,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 2.003157138824463,
+ "step": 1855
+ },
+ {
+ "epoch": 25.78048780487805,
+ "grad_norm": 0.0612902007997036,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 1.9832231998443604,
+ "step": 1856
+ },
+ {
+ "epoch": 25.794425087108014,
+ "grad_norm": 0.0625016912817955,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 2.0012407302856445,
+ "step": 1857
+ },
+ {
+ "epoch": 25.808362369337978,
+ "grad_norm": 0.062357913702726364,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 2.010338306427002,
+ "step": 1858
+ },
+ {
+ "epoch": 25.822299651567945,
+ "grad_norm": 0.06349784880876541,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 2.0105648040771484,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83623693379791,
+ "grad_norm": 0.061822179704904556,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 1.9987019300460815,
+ "step": 1860
+ },
+ {
+ "epoch": 25.850174216027874,
+ "grad_norm": 0.06443580985069275,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 1.994943618774414,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86411149825784,
+ "grad_norm": 0.061674103140830994,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 2.0116522312164307,
+ "step": 1862
+ },
+ {
+ "epoch": 25.878048780487806,
+ "grad_norm": 0.061944279819726944,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 2.012296438217163,
+ "step": 1863
+ },
+ {
+ "epoch": 25.89198606271777,
+ "grad_norm": 0.06270956248044968,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 2.0057339668273926,
+ "step": 1864
+ },
+ {
+ "epoch": 25.905923344947734,
+ "grad_norm": 0.06427973508834839,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 2.0053863525390625,
+ "step": 1865
+ },
+ {
+ "epoch": 25.9198606271777,
+ "grad_norm": 0.0637369304895401,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 2.0130667686462402,
+ "step": 1866
+ },
+ {
+ "epoch": 25.933797909407666,
+ "grad_norm": 0.06291835755109787,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 2.0094404220581055,
+ "step": 1867
+ },
+ {
+ "epoch": 25.94773519163763,
+ "grad_norm": 0.06374962627887726,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 2.0074503421783447,
+ "step": 1868
+ },
+ {
+ "epoch": 25.961672473867594,
+ "grad_norm": 0.06330475956201553,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 1.9999325275421143,
+ "step": 1869
+ },
+ {
+ "epoch": 25.975609756097562,
+ "grad_norm": 0.06077918782830238,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 1.994345784187317,
+ "step": 1870
+ },
+ {
+ "epoch": 25.989547038327526,
+ "grad_norm": 0.06524437665939331,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 2.0210394859313965,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.05064176768064499,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 1.4825201034545898,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6298588514328003,
+ "eval_runtime": 43.1768,
+ "eval_samples_per_second": 56.558,
+ "eval_steps_per_second": 0.463,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013937282229964,
+ "grad_norm": 0.05861504375934601,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 1.9728282690048218,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027874564459932,
+ "grad_norm": 0.06309985369443893,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 1.9638245105743408,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041811846689896,
+ "grad_norm": 0.06438111513853073,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 1.9697308540344238,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05574912891986,
+ "grad_norm": 0.060332708060741425,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 1.9574062824249268,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069686411149824,
+ "grad_norm": 0.061463311314582825,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 1.9469952583312988,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083623693379792,
+ "grad_norm": 0.06337480992078781,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 1.9695435762405396,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097560975609756,
+ "grad_norm": 0.06303273886442184,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 1.9764964580535889,
+ "step": 1879
+ },
+ {
+ "epoch": 26.11149825783972,
+ "grad_norm": 0.06061594933271408,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 1.9758808612823486,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125435540069688,
+ "grad_norm": 0.061181873083114624,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 1.9607892036437988,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139372822299652,
+ "grad_norm": 0.06557024270296097,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 1.9752333164215088,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153310104529616,
+ "grad_norm": 0.06261707097291946,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 1.9634147882461548,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16724738675958,
+ "grad_norm": 0.05993996560573578,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 1.957287311553955,
+ "step": 1884
+ },
+ {
+ "epoch": 26.181184668989548,
+ "grad_norm": 0.0612766332924366,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 1.9562265872955322,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195121951219512,
+ "grad_norm": 0.05735815688967705,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 1.9577300548553467,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209059233449477,
+ "grad_norm": 0.06258611381053925,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 1.9661660194396973,
+ "step": 1887
+ },
+ {
+ "epoch": 26.222996515679444,
+ "grad_norm": 0.06298062205314636,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 1.9508070945739746,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23693379790941,
+ "grad_norm": 0.06346822530031204,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 1.9677343368530273,
+ "step": 1889
+ },
+ {
+ "epoch": 26.250871080139373,
+ "grad_norm": 0.05959425866603851,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 1.9574894905090332,
+ "step": 1890
+ },
+ {
+ "epoch": 26.264808362369337,
+ "grad_norm": 0.0610729418694973,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 1.9515372514724731,
+ "step": 1891
+ },
+ {
+ "epoch": 26.278745644599304,
+ "grad_norm": 0.06362602859735489,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 1.966064453125,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29268292682927,
+ "grad_norm": 0.06124011054635048,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 1.9654541015625,
+ "step": 1893
+ },
+ {
+ "epoch": 26.306620209059233,
+ "grad_norm": 0.05773317068815231,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 1.9610934257507324,
+ "step": 1894
+ },
+ {
+ "epoch": 26.320557491289197,
+ "grad_norm": 0.05896785855293274,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 1.965133547782898,
+ "step": 1895
+ },
+ {
+ "epoch": 26.334494773519165,
+ "grad_norm": 0.059504859149456024,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 1.9617180824279785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.34843205574913,
+ "grad_norm": 0.05804497376084328,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 1.9761323928833008,
+ "step": 1897
+ },
+ {
+ "epoch": 26.362369337979093,
+ "grad_norm": 0.058211296796798706,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 1.9805936813354492,
+ "step": 1898
+ },
+ {
+ "epoch": 26.37630662020906,
+ "grad_norm": 0.05976752191781998,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 1.9514508247375488,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390243902439025,
+ "grad_norm": 0.05824121832847595,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 1.97285795211792,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40418118466899,
+ "grad_norm": 0.05793534219264984,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 1.9608209133148193,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418118466898953,
+ "grad_norm": 0.05697625130414963,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 1.9787089824676514,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43205574912892,
+ "grad_norm": 0.05703677609562874,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 1.9610021114349365,
+ "step": 1903
+ },
+ {
+ "epoch": 26.445993031358885,
+ "grad_norm": 0.057380311191082,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 1.9764081239700317,
+ "step": 1904
+ },
+ {
+ "epoch": 26.45993031358885,
+ "grad_norm": 0.05846763774752617,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 1.9641292095184326,
+ "step": 1905
+ },
+ {
+ "epoch": 26.473867595818817,
+ "grad_norm": 0.05929148569703102,
+ "learning_rate": 0.000288343693342466,
+ "loss": 1.9803684949874878,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48780487804878,
+ "grad_norm": 0.058236971497535706,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 1.981757402420044,
+ "step": 1907
+ },
+ {
+ "epoch": 26.501742160278745,
+ "grad_norm": 0.05845103785395622,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 1.9736125469207764,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51567944250871,
+ "grad_norm": 0.06104041263461113,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 1.9698188304901123,
+ "step": 1909
+ },
+ {
+ "epoch": 26.529616724738677,
+ "grad_norm": 0.0628688707947731,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 1.9786568880081177,
+ "step": 1910
+ },
+ {
+ "epoch": 26.54355400696864,
+ "grad_norm": 0.06115978956222534,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 1.9744799137115479,
+ "step": 1911
+ },
+ {
+ "epoch": 26.557491289198605,
+ "grad_norm": 0.06233362853527069,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 1.9801995754241943,
+ "step": 1912
+ },
+ {
+ "epoch": 26.571428571428573,
+ "grad_norm": 0.06182500347495079,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 1.9888432025909424,
+ "step": 1913
+ },
+ {
+ "epoch": 26.585365853658537,
+ "grad_norm": 0.05962767079472542,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 1.9648957252502441,
+ "step": 1914
+ },
+ {
+ "epoch": 26.5993031358885,
+ "grad_norm": 0.05994252488017082,
+ "learning_rate": 0.000285947841605349,
+ "loss": 1.9709041118621826,
+ "step": 1915
+ },
+ {
+ "epoch": 26.613240418118465,
+ "grad_norm": 0.06045769527554512,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 1.994588017463684,
+ "step": 1916
+ },
+ {
+ "epoch": 26.627177700348433,
+ "grad_norm": 0.059465229511260986,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 1.9743266105651855,
+ "step": 1917
+ },
+ {
+ "epoch": 26.641114982578397,
+ "grad_norm": 0.06020190939307213,
+ "learning_rate": 0.000285149463934261,
+ "loss": 1.9780995845794678,
+ "step": 1918
+ },
+ {
+ "epoch": 26.65505226480836,
+ "grad_norm": 0.05829813703894615,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 1.9831392765045166,
+ "step": 1919
+ },
+ {
+ "epoch": 26.66898954703833,
+ "grad_norm": 0.05960174649953842,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 1.9799871444702148,
+ "step": 1920
+ },
+ {
+ "epoch": 26.682926829268293,
+ "grad_norm": 0.06133116036653519,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 1.985158920288086,
+ "step": 1921
+ },
+ {
+ "epoch": 26.696864111498257,
+ "grad_norm": 0.05947015434503555,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 1.9783923625946045,
+ "step": 1922
+ },
+ {
+ "epoch": 26.71080139372822,
+ "grad_norm": 0.061687882989645004,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 1.9685261249542236,
+ "step": 1923
+ },
+ {
+ "epoch": 26.72473867595819,
+ "grad_norm": 0.0613567978143692,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 1.9945828914642334,
+ "step": 1924
+ },
+ {
+ "epoch": 26.738675958188153,
+ "grad_norm": 0.05902605876326561,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 1.966120719909668,
+ "step": 1925
+ },
+ {
+ "epoch": 26.752613240418118,
+ "grad_norm": 0.06018664687871933,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 1.991615653038025,
+ "step": 1926
+ },
+ {
+ "epoch": 26.766550522648085,
+ "grad_norm": 0.061171986162662506,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 1.9877536296844482,
+ "step": 1927
+ },
+ {
+ "epoch": 26.78048780487805,
+ "grad_norm": 0.06263386458158493,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 1.9799100160598755,
+ "step": 1928
+ },
+ {
+ "epoch": 26.794425087108014,
+ "grad_norm": 0.06326548755168915,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 1.9805757999420166,
+ "step": 1929
+ },
+ {
+ "epoch": 26.808362369337978,
+ "grad_norm": 0.06226228550076485,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 1.9867990016937256,
+ "step": 1930
+ },
+ {
+ "epoch": 26.822299651567945,
+ "grad_norm": 0.06003205478191376,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 1.9632174968719482,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83623693379791,
+ "grad_norm": 0.06075739488005638,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 1.9638378620147705,
+ "step": 1932
+ },
+ {
+ "epoch": 26.850174216027874,
+ "grad_norm": 0.061726443469524384,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 1.9865589141845703,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86411149825784,
+ "grad_norm": 0.0621049702167511,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 2.007071018218994,
+ "step": 1934
+ },
+ {
+ "epoch": 26.878048780487806,
+ "grad_norm": 0.061793744564056396,
+ "learning_rate": 0.000280627938758204,
+ "loss": 1.988924264907837,
+ "step": 1935
+ },
+ {
+ "epoch": 26.89198606271777,
+ "grad_norm": 0.062355563044548035,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 1.9830653667449951,
+ "step": 1936
+ },
+ {
+ "epoch": 26.905923344947734,
+ "grad_norm": 0.060881514102220535,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 2.003551959991455,
+ "step": 1937
+ },
+ {
+ "epoch": 26.9198606271777,
+ "grad_norm": 0.061703141778707504,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 1.9811534881591797,
+ "step": 1938
+ },
+ {
+ "epoch": 26.933797909407666,
+ "grad_norm": 0.0634910985827446,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 1.9840846061706543,
+ "step": 1939
+ },
+ {
+ "epoch": 26.94773519163763,
+ "grad_norm": 0.060807306319475174,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 1.9879541397094727,
+ "step": 1940
+ },
+ {
+ "epoch": 26.961672473867594,
+ "grad_norm": 0.06126011535525322,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 1.993988037109375,
+ "step": 1941
+ },
+ {
+ "epoch": 26.975609756097562,
+ "grad_norm": 0.06394185870885849,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 1.983365535736084,
+ "step": 1942
+ },
+ {
+ "epoch": 26.989547038327526,
+ "grad_norm": 0.06164516508579254,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 1.989990234375,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.04997039958834648,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 1.4921388626098633,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6310494542121887,
+ "eval_runtime": 40.875,
+ "eval_samples_per_second": 59.743,
+ "eval_steps_per_second": 0.489,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013937282229964,
+ "grad_norm": 0.05894223228096962,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 1.935828447341919,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027874564459932,
+ "grad_norm": 0.06731119751930237,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 1.945769190788269,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041811846689896,
+ "grad_norm": 0.06787799298763275,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 1.944061040878296,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05574912891986,
+ "grad_norm": 0.05840576812624931,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 1.94246506690979,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069686411149824,
+ "grad_norm": 0.058151841163635254,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 1.940029263496399,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083623693379792,
+ "grad_norm": 0.06329379975795746,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 1.9523427486419678,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097560975609756,
+ "grad_norm": 0.06530784070491791,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 1.9352362155914307,
+ "step": 1951
+ },
+ {
+ "epoch": 27.11149825783972,
+ "grad_norm": 0.06273649632930756,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 1.946252703666687,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125435540069688,
+ "grad_norm": 0.05996445193886757,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 1.9437150955200195,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139372822299652,
+ "grad_norm": 0.059087123721838,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 1.936135172843933,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153310104529616,
+ "grad_norm": 0.06257616728544235,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 1.9452544450759888,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16724738675958,
+ "grad_norm": 0.06591545045375824,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 1.9334893226623535,
+ "step": 1956
+ },
+ {
+ "epoch": 27.181184668989548,
+ "grad_norm": 0.0595872700214386,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 1.9514997005462646,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195121951219512,
+ "grad_norm": 0.05855429172515869,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 1.955148458480835,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209059233449477,
+ "grad_norm": 0.06383495032787323,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 1.9607688188552856,
+ "step": 1959
+ },
+ {
+ "epoch": 27.222996515679444,
+ "grad_norm": 0.06486682593822479,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 1.941009283065796,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23693379790941,
+ "grad_norm": 0.06293730437755585,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 1.9550879001617432,
+ "step": 1961
+ },
+ {
+ "epoch": 27.250871080139373,
+ "grad_norm": 0.06104227155447006,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 1.9483157396316528,
+ "step": 1962
+ },
+ {
+ "epoch": 27.264808362369337,
+ "grad_norm": 0.06920639425516129,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 1.948688268661499,
+ "step": 1963
+ },
+ {
+ "epoch": 27.278745644599304,
+ "grad_norm": 0.06186635419726372,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 1.9512240886688232,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29268292682927,
+ "grad_norm": 0.06258274614810944,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 1.9538028240203857,
+ "step": 1965
+ },
+ {
+ "epoch": 27.306620209059233,
+ "grad_norm": 0.06546445935964584,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 1.9497663974761963,
+ "step": 1966
+ },
+ {
+ "epoch": 27.320557491289197,
+ "grad_norm": 0.06289096921682358,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 1.9706182479858398,
+ "step": 1967
+ },
+ {
+ "epoch": 27.334494773519165,
+ "grad_norm": 0.06412214040756226,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 1.9423331022262573,
+ "step": 1968
+ },
+ {
+ "epoch": 27.34843205574913,
+ "grad_norm": 0.062773697078228,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 1.9471882581710815,
+ "step": 1969
+ },
+ {
+ "epoch": 27.362369337979093,
+ "grad_norm": 0.060290850698947906,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 1.9732742309570312,
+ "step": 1970
+ },
+ {
+ "epoch": 27.37630662020906,
+ "grad_norm": 0.062357380986213684,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 1.957550048828125,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390243902439025,
+ "grad_norm": 0.06070727854967117,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 1.951438069343567,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40418118466899,
+ "grad_norm": 0.06380726397037506,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 1.9512609243392944,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418118466898953,
+ "grad_norm": 0.06135795637965202,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 1.9663273096084595,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43205574912892,
+ "grad_norm": 0.0627567246556282,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 1.964203119277954,
+ "step": 1975
+ },
+ {
+ "epoch": 27.445993031358885,
+ "grad_norm": 0.06235141307115555,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 1.960247278213501,
+ "step": 1976
+ },
+ {
+ "epoch": 27.45993031358885,
+ "grad_norm": 0.06187676638364792,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 1.9667054414749146,
+ "step": 1977
+ },
+ {
+ "epoch": 27.473867595818817,
+ "grad_norm": 0.06090894341468811,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 1.9695777893066406,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48780487804878,
+ "grad_norm": 0.06363680213689804,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 1.9529361724853516,
+ "step": 1979
+ },
+ {
+ "epoch": 27.501742160278745,
+ "grad_norm": 0.06053898110985756,
+ "learning_rate": 0.00026868712586269,
+ "loss": 1.9707324504852295,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51567944250871,
+ "grad_norm": 0.061363156884908676,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 1.955542802810669,
+ "step": 1981
+ },
+ {
+ "epoch": 27.529616724738677,
+ "grad_norm": 0.06158224865794182,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 1.9489967823028564,
+ "step": 1982
+ },
+ {
+ "epoch": 27.54355400696864,
+ "grad_norm": 0.05886901170015335,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 1.9705257415771484,
+ "step": 1983
+ },
+ {
+ "epoch": 27.557491289198605,
+ "grad_norm": 0.05890485271811485,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 1.9430618286132812,
+ "step": 1984
+ },
+ {
+ "epoch": 27.571428571428573,
+ "grad_norm": 0.060433268547058105,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 1.9634191989898682,
+ "step": 1985
+ },
+ {
+ "epoch": 27.585365853658537,
+ "grad_norm": 0.0626354068517685,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 1.9568289518356323,
+ "step": 1986
+ },
+ {
+ "epoch": 27.5993031358885,
+ "grad_norm": 0.05973317474126816,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 1.9678292274475098,
+ "step": 1987
+ },
+ {
+ "epoch": 27.613240418118465,
+ "grad_norm": 0.062289632856845856,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 1.9413363933563232,
+ "step": 1988
+ },
+ {
+ "epoch": 27.627177700348433,
+ "grad_norm": 0.06162180379033089,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 1.9835751056671143,
+ "step": 1989
+ },
+ {
+ "epoch": 27.641114982578397,
+ "grad_norm": 0.06176188960671425,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 1.9541220664978027,
+ "step": 1990
+ },
+ {
+ "epoch": 27.65505226480836,
+ "grad_norm": 0.06479815393686295,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 1.9596972465515137,
+ "step": 1991
+ },
+ {
+ "epoch": 27.66898954703833,
+ "grad_norm": 0.06571854650974274,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 1.9628980159759521,
+ "step": 1992
+ },
+ {
+ "epoch": 27.682926829268293,
+ "grad_norm": 0.06167568638920784,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 1.9624769687652588,
+ "step": 1993
+ },
+ {
+ "epoch": 27.696864111498257,
+ "grad_norm": 0.06207694113254547,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 1.9487217664718628,
+ "step": 1994
+ },
+ {
+ "epoch": 27.71080139372822,
+ "grad_norm": 0.061430424451828,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 1.9567784070968628,
+ "step": 1995
+ },
+ {
+ "epoch": 27.72473867595819,
+ "grad_norm": 0.06589192897081375,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 1.9594354629516602,
+ "step": 1996
+ },
+ {
+ "epoch": 27.738675958188153,
+ "grad_norm": 0.06640033423900604,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 1.9637835025787354,
+ "step": 1997
+ },
+ {
+ "epoch": 27.752613240418118,
+ "grad_norm": 0.06181980296969414,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 1.9597129821777344,
+ "step": 1998
+ },
+ {
+ "epoch": 27.766550522648085,
+ "grad_norm": 0.06352445483207703,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 1.9612562656402588,
+ "step": 1999
+ },
+ {
+ "epoch": 27.78048780487805,
+ "grad_norm": 0.06398574262857437,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 1.9662930965423584,
+ "step": 2000
+ },
+ {
+ "epoch": 27.794425087108014,
+ "grad_norm": 0.06195084750652313,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 1.956035852432251,
+ "step": 2001
+ },
+ {
+ "epoch": 27.808362369337978,
+ "grad_norm": 0.05951351299881935,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 1.9712438583374023,
+ "step": 2002
+ },
+ {
+ "epoch": 27.822299651567945,
+ "grad_norm": 0.062185633927583694,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 1.9714539051055908,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83623693379791,
+ "grad_norm": 0.061178892850875854,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 1.9691438674926758,
+ "step": 2004
+ },
+ {
+ "epoch": 27.850174216027874,
+ "grad_norm": 0.06262431293725967,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 1.9506231546401978,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86411149825784,
+ "grad_norm": 0.06206081807613373,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 1.9662809371948242,
+ "step": 2006
+ },
+ {
+ "epoch": 27.878048780487806,
+ "grad_norm": 0.0626990869641304,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 1.972696304321289,
+ "step": 2007
+ },
+ {
+ "epoch": 27.89198606271777,
+ "grad_norm": 0.06345485150814056,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 1.9560344219207764,
+ "step": 2008
+ },
+ {
+ "epoch": 27.905923344947734,
+ "grad_norm": 0.06401040405035019,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 1.9732205867767334,
+ "step": 2009
+ },
+ {
+ "epoch": 27.9198606271777,
+ "grad_norm": 0.06437943130731583,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 1.974897027015686,
+ "step": 2010
+ },
+ {
+ "epoch": 27.933797909407666,
+ "grad_norm": 0.06296383589506149,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 1.9710102081298828,
+ "step": 2011
+ },
+ {
+ "epoch": 27.94773519163763,
+ "grad_norm": 0.06527025997638702,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 1.9707660675048828,
+ "step": 2012
+ },
+ {
+ "epoch": 27.961672473867594,
+ "grad_norm": 0.06185369938611984,
+ "learning_rate": 0.00025996500713765,
+ "loss": 1.9675114154815674,
+ "step": 2013
+ },
+ {
+ "epoch": 27.975609756097562,
+ "grad_norm": 0.06357269734144211,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 1.9812211990356445,
+ "step": 2014
+ },
+ {
+ "epoch": 27.989547038327526,
+ "grad_norm": 0.06581240892410278,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 1.97153639793396,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.049184102565050125,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 1.4647765159606934,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6328383088111877,
+ "eval_runtime": 40.5927,
+ "eval_samples_per_second": 60.159,
+ "eval_steps_per_second": 0.493,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013937282229964,
+ "grad_norm": 0.05999046564102173,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 1.9182333946228027,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027874564459932,
+ "grad_norm": 0.059804853051900864,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 1.924762487411499,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041811846689896,
+ "grad_norm": 0.05890650674700737,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 1.9341628551483154,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05574912891986,
+ "grad_norm": 0.057233694940805435,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 1.932476282119751,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069686411149824,
+ "grad_norm": 0.060101717710494995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 1.9403120279312134,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083623693379792,
+ "grad_norm": 0.06054912135004997,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 1.9274934530258179,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097560975609756,
+ "grad_norm": 0.05882031098008156,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 1.9331915378570557,
+ "step": 2023
+ },
+ {
+ "epoch": 28.11149825783972,
+ "grad_norm": 0.05846688151359558,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 1.9342529773712158,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125435540069688,
+ "grad_norm": 0.058974213898181915,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 1.9353152513504028,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139372822299652,
+ "grad_norm": 0.05882132798433304,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 1.9500072002410889,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153310104529616,
+ "grad_norm": 0.05919627472758293,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 1.9313960075378418,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16724738675958,
+ "grad_norm": 0.06413958966732025,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 1.9516873359680176,
+ "step": 2028
+ },
+ {
+ "epoch": 28.181184668989548,
+ "grad_norm": 0.05799189209938049,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 1.9290452003479004,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195121951219512,
+ "grad_norm": 0.05999152734875679,
+ "learning_rate": 0.000255486047794226,
+ "loss": 1.9396628141403198,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209059233449477,
+ "grad_norm": 0.060240138322114944,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 1.937988519668579,
+ "step": 2031
+ },
+ {
+ "epoch": 28.222996515679444,
+ "grad_norm": 0.05933740735054016,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 1.9343246221542358,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23693379790941,
+ "grad_norm": 0.05960093438625336,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 1.9326903820037842,
+ "step": 2033
+ },
+ {
+ "epoch": 28.250871080139373,
+ "grad_norm": 0.0635334923863411,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 1.9381543397903442,
+ "step": 2034
+ },
+ {
+ "epoch": 28.264808362369337,
+ "grad_norm": 0.06162877380847931,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 1.9265705347061157,
+ "step": 2035
+ },
+ {
+ "epoch": 28.278745644599304,
+ "grad_norm": 0.059690047055482864,
+ "learning_rate": 0.000253907826333243,
+ "loss": 1.9278099536895752,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29268292682927,
+ "grad_norm": 0.0612013153731823,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 1.939026117324829,
+ "step": 2037
+ },
+ {
+ "epoch": 28.306620209059233,
+ "grad_norm": 0.06080419942736626,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 1.9306461811065674,
+ "step": 2038
+ },
+ {
+ "epoch": 28.320557491289197,
+ "grad_norm": 0.06308622658252716,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 1.9350788593292236,
+ "step": 2039
+ },
+ {
+ "epoch": 28.334494773519165,
+ "grad_norm": 0.06124545633792877,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 1.9264733791351318,
+ "step": 2040
+ },
+ {
+ "epoch": 28.34843205574913,
+ "grad_norm": 0.059037938714027405,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 1.917725920677185,
+ "step": 2041
+ },
+ {
+ "epoch": 28.362369337979093,
+ "grad_norm": 0.06089222431182861,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 1.949327826499939,
+ "step": 2042
+ },
+ {
+ "epoch": 28.37630662020906,
+ "grad_norm": 0.05957084521651268,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 1.938978910446167,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390243902439025,
+ "grad_norm": 0.06060048192739487,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 1.9226362705230713,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40418118466899,
+ "grad_norm": 0.058304425328969955,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 1.9330644607543945,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418118466898953,
+ "grad_norm": 0.05970550701022148,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 1.931627631187439,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43205574912892,
+ "grad_norm": 0.05994785949587822,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 1.9387202262878418,
+ "step": 2047
+ },
+ {
+ "epoch": 28.445993031358885,
+ "grad_norm": 0.06001518666744232,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 1.951300859451294,
+ "step": 2048
+ },
+ {
+ "epoch": 28.45993031358885,
+ "grad_norm": 0.0636107549071312,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 1.9358865022659302,
+ "step": 2049
+ },
+ {
+ "epoch": 28.473867595818817,
+ "grad_norm": 0.059402115643024445,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 1.9298653602600098,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48780487804878,
+ "grad_norm": 0.06260179728269577,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 1.954633116722107,
+ "step": 2051
+ },
+ {
+ "epoch": 28.501742160278745,
+ "grad_norm": 0.06439191848039627,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 1.9338306188583374,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51567944250871,
+ "grad_norm": 0.06265310943126678,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 1.9500043392181396,
+ "step": 2053
+ },
+ {
+ "epoch": 28.529616724738677,
+ "grad_norm": 0.06323659420013428,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 1.9445586204528809,
+ "step": 2054
+ },
+ {
+ "epoch": 28.54355400696864,
+ "grad_norm": 0.06512445956468582,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 1.958587408065796,
+ "step": 2055
+ },
+ {
+ "epoch": 28.557491289198605,
+ "grad_norm": 0.06386571377515793,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 1.949887990951538,
+ "step": 2056
+ },
+ {
+ "epoch": 28.571428571428573,
+ "grad_norm": 0.06317705661058426,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 1.9512674808502197,
+ "step": 2057
+ },
+ {
+ "epoch": 28.585365853658537,
+ "grad_norm": 0.06504984200000763,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 1.9242960214614868,
+ "step": 2058
+ },
+ {
+ "epoch": 28.5993031358885,
+ "grad_norm": 0.06278593838214874,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 1.9203131198883057,
+ "step": 2059
+ },
+ {
+ "epoch": 28.613240418118465,
+ "grad_norm": 0.06171257048845291,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 1.954156517982483,
+ "step": 2060
+ },
+ {
+ "epoch": 28.627177700348433,
+ "grad_norm": 0.06303241103887558,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 1.9460458755493164,
+ "step": 2061
+ },
+ {
+ "epoch": 28.641114982578397,
+ "grad_norm": 0.06252635270357132,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 1.9477293491363525,
+ "step": 2062
+ },
+ {
+ "epoch": 28.65505226480836,
+ "grad_norm": 0.06226412579417229,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 1.9259647130966187,
+ "step": 2063
+ },
+ {
+ "epoch": 28.66898954703833,
+ "grad_norm": 0.06111626327037811,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 1.9573581218719482,
+ "step": 2064
+ },
+ {
+ "epoch": 28.682926829268293,
+ "grad_norm": 0.06460461020469666,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 1.9486393928527832,
+ "step": 2065
+ },
+ {
+ "epoch": 28.696864111498257,
+ "grad_norm": 0.062027834355831146,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 1.9404425621032715,
+ "step": 2066
+ },
+ {
+ "epoch": 28.71080139372822,
+ "grad_norm": 0.061942074447870255,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 1.9444060325622559,
+ "step": 2067
+ },
+ {
+ "epoch": 28.72473867595819,
+ "grad_norm": 0.06159934028983116,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 1.9635517597198486,
+ "step": 2068
+ },
+ {
+ "epoch": 28.738675958188153,
+ "grad_norm": 0.0621839202940464,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 1.9442765712738037,
+ "step": 2069
+ },
+ {
+ "epoch": 28.752613240418118,
+ "grad_norm": 0.06432677060365677,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 1.9604840278625488,
+ "step": 2070
+ },
+ {
+ "epoch": 28.766550522648085,
+ "grad_norm": 0.06037316843867302,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 1.948632836341858,
+ "step": 2071
+ },
+ {
+ "epoch": 28.78048780487805,
+ "grad_norm": 0.05999939888715744,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 1.9451751708984375,
+ "step": 2072
+ },
+ {
+ "epoch": 28.794425087108014,
+ "grad_norm": 0.06441698968410492,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 1.9356517791748047,
+ "step": 2073
+ },
+ {
+ "epoch": 28.808362369337978,
+ "grad_norm": 0.06059470772743225,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 1.9258331060409546,
+ "step": 2074
+ },
+ {
+ "epoch": 28.822299651567945,
+ "grad_norm": 0.062317874282598495,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 1.9342000484466553,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83623693379791,
+ "grad_norm": 0.06318258494138718,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 1.9395548105239868,
+ "step": 2076
+ },
+ {
+ "epoch": 28.850174216027874,
+ "grad_norm": 0.06204817444086075,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 1.9572374820709229,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86411149825784,
+ "grad_norm": 0.06250188499689102,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 1.9465739727020264,
+ "step": 2078
+ },
+ {
+ "epoch": 28.878048780487806,
+ "grad_norm": 0.06293781846761703,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 1.9431169033050537,
+ "step": 2079
+ },
+ {
+ "epoch": 28.89198606271777,
+ "grad_norm": 0.059349458664655685,
+ "learning_rate": 0.000242380656502223,
+ "loss": 1.9673197269439697,
+ "step": 2080
+ },
+ {
+ "epoch": 28.905923344947734,
+ "grad_norm": 0.06214660778641701,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 1.9646787643432617,
+ "step": 2081
+ },
+ {
+ "epoch": 28.9198606271777,
+ "grad_norm": 0.062279123812913895,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 1.9572803974151611,
+ "step": 2082
+ },
+ {
+ "epoch": 28.933797909407666,
+ "grad_norm": 0.06502588093280792,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 1.9510462284088135,
+ "step": 2083
+ },
+ {
+ "epoch": 28.94773519163763,
+ "grad_norm": 0.06153101101517677,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 1.9507923126220703,
+ "step": 2084
+ },
+ {
+ "epoch": 28.961672473867594,
+ "grad_norm": 0.06469229608774185,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 1.937751293182373,
+ "step": 2085
+ },
+ {
+ "epoch": 28.975609756097562,
+ "grad_norm": 0.0672568678855896,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 1.944124698638916,
+ "step": 2086
+ },
+ {
+ "epoch": 28.989547038327526,
+ "grad_norm": 0.06451335549354553,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 1.9519702196121216,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.05141456052660942,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 1.4445865154266357,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6343450546264648,
+ "eval_runtime": 40.5081,
+ "eval_samples_per_second": 60.284,
+ "eval_steps_per_second": 0.494,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013937282229964,
+ "grad_norm": 0.061562713235616684,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 1.921946406364441,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027874564459932,
+ "grad_norm": 0.06396839767694473,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 1.9222416877746582,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041811846689896,
+ "grad_norm": 0.05828692391514778,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 1.9173905849456787,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05574912891986,
+ "grad_norm": 0.06567556411027908,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 1.9148579835891724,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069686411149824,
+ "grad_norm": 0.05924173817038536,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 1.916333794593811,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083623693379792,
+ "grad_norm": 0.06129243224859238,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 1.9031038284301758,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097560975609756,
+ "grad_norm": 0.06068028509616852,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 1.9165983200073242,
+ "step": 2095
+ },
+ {
+ "epoch": 29.11149825783972,
+ "grad_norm": 0.06304074823856354,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 1.90432870388031,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125435540069688,
+ "grad_norm": 0.060179345309734344,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 1.9187151193618774,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139372822299652,
+ "grad_norm": 0.05633116513490677,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 1.9125087261199951,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153310104529616,
+ "grad_norm": 0.0598532073199749,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 1.906031847000122,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16724738675958,
+ "grad_norm": 0.05987119674682617,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 1.9177607297897339,
+ "step": 2100
+ },
+ {
+ "epoch": 29.181184668989548,
+ "grad_norm": 0.060883522033691406,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 1.9008433818817139,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195121951219512,
+ "grad_norm": 0.05977628752589226,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 1.9176266193389893,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209059233449477,
+ "grad_norm": 0.05939047038555145,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 1.9227800369262695,
+ "step": 2103
+ },
+ {
+ "epoch": 29.222996515679444,
+ "grad_norm": 0.059280600398778915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 1.9164855480194092,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23693379790941,
+ "grad_norm": 0.058746229857206345,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 1.9110995531082153,
+ "step": 2105
+ },
+ {
+ "epoch": 29.250871080139373,
+ "grad_norm": 0.05882417410612106,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 1.926957368850708,
+ "step": 2106
+ },
+ {
+ "epoch": 29.264808362369337,
+ "grad_norm": 0.05827188491821289,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 1.9101874828338623,
+ "step": 2107
+ },
+ {
+ "epoch": 29.278745644599304,
+ "grad_norm": 0.059739116579294205,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 1.924944519996643,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29268292682927,
+ "grad_norm": 0.058013755828142166,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 1.9121344089508057,
+ "step": 2109
+ },
+ {
+ "epoch": 29.306620209059233,
+ "grad_norm": 0.061207350343465805,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 1.9137616157531738,
+ "step": 2110
+ },
+ {
+ "epoch": 29.320557491289197,
+ "grad_norm": 0.05848725140094757,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 1.9290939569473267,
+ "step": 2111
+ },
+ {
+ "epoch": 29.334494773519165,
+ "grad_norm": 0.059134867042303085,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 1.919262409210205,
+ "step": 2112
+ },
+ {
+ "epoch": 29.34843205574913,
+ "grad_norm": 0.06025470420718193,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 1.9021251201629639,
+ "step": 2113
+ },
+ {
+ "epoch": 29.362369337979093,
+ "grad_norm": 0.05981536582112312,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 1.9163634777069092,
+ "step": 2114
+ },
+ {
+ "epoch": 29.37630662020906,
+ "grad_norm": 0.05826934054493904,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 1.8987491130828857,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390243902439025,
+ "grad_norm": 0.06095297262072563,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 1.9322285652160645,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40418118466899,
+ "grad_norm": 0.05814153701066971,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 1.927016019821167,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418118466898953,
+ "grad_norm": 0.05872226879000664,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 1.9271752834320068,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43205574912892,
+ "grad_norm": 0.06008293107151985,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 1.9269304275512695,
+ "step": 2119
+ },
+ {
+ "epoch": 29.445993031358885,
+ "grad_norm": 0.06035081297159195,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 1.929135799407959,
+ "step": 2120
+ },
+ {
+ "epoch": 29.45993031358885,
+ "grad_norm": 0.06080561876296997,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 1.9284312725067139,
+ "step": 2121
+ },
+ {
+ "epoch": 29.473867595818817,
+ "grad_norm": 0.058793697506189346,
+ "learning_rate": 0.000231465389734324,
+ "loss": 1.931719183921814,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48780487804878,
+ "grad_norm": 0.06385456770658493,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 1.9112051725387573,
+ "step": 2123
+ },
+ {
+ "epoch": 29.501742160278745,
+ "grad_norm": 0.06030426174402237,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 1.9254870414733887,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51567944250871,
+ "grad_norm": 0.059933315962553024,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 1.915220022201538,
+ "step": 2125
+ },
+ {
+ "epoch": 29.529616724738677,
+ "grad_norm": 0.06083494424819946,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 1.9391965866088867,
+ "step": 2126
+ },
+ {
+ "epoch": 29.54355400696864,
+ "grad_norm": 0.06049460917711258,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 1.9301774501800537,
+ "step": 2127
+ },
+ {
+ "epoch": 29.557491289198605,
+ "grad_norm": 0.05953316017985344,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 1.9150071144104004,
+ "step": 2128
+ },
+ {
+ "epoch": 29.571428571428573,
+ "grad_norm": 0.0611971914768219,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 1.9265213012695312,
+ "step": 2129
+ },
+ {
+ "epoch": 29.585365853658537,
+ "grad_norm": 0.06123765930533409,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 1.9405107498168945,
+ "step": 2130
+ },
+ {
+ "epoch": 29.5993031358885,
+ "grad_norm": 0.06403640657663345,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 1.9216606616973877,
+ "step": 2131
+ },
+ {
+ "epoch": 29.613240418118465,
+ "grad_norm": 0.06530711054801941,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 1.9302153587341309,
+ "step": 2132
+ },
+ {
+ "epoch": 29.627177700348433,
+ "grad_norm": 0.06106860563158989,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 1.930979609489441,
+ "step": 2133
+ },
+ {
+ "epoch": 29.641114982578397,
+ "grad_norm": 0.06148209050297737,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 1.9276363849639893,
+ "step": 2134
+ },
+ {
+ "epoch": 29.65505226480836,
+ "grad_norm": 0.060283321887254715,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 1.9269459247589111,
+ "step": 2135
+ },
+ {
+ "epoch": 29.66898954703833,
+ "grad_norm": 0.06138703227043152,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 1.9389945268630981,
+ "step": 2136
+ },
+ {
+ "epoch": 29.682926829268293,
+ "grad_norm": 0.06101245433092117,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 1.9426372051239014,
+ "step": 2137
+ },
+ {
+ "epoch": 29.696864111498257,
+ "grad_norm": 0.05958810821175575,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 1.9327162504196167,
+ "step": 2138
+ },
+ {
+ "epoch": 29.71080139372822,
+ "grad_norm": 0.06250876933336258,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 1.9384305477142334,
+ "step": 2139
+ },
+ {
+ "epoch": 29.72473867595819,
+ "grad_norm": 0.06344381719827652,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 1.9422640800476074,
+ "step": 2140
+ },
+ {
+ "epoch": 29.738675958188153,
+ "grad_norm": 0.06065250560641289,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 1.920068621635437,
+ "step": 2141
+ },
+ {
+ "epoch": 29.752613240418118,
+ "grad_norm": 0.06333960592746735,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 1.925317645072937,
+ "step": 2142
+ },
+ {
+ "epoch": 29.766550522648085,
+ "grad_norm": 0.06285370141267776,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 1.92912757396698,
+ "step": 2143
+ },
+ {
+ "epoch": 29.78048780487805,
+ "grad_norm": 0.06441807746887207,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 1.941272497177124,
+ "step": 2144
+ },
+ {
+ "epoch": 29.794425087108014,
+ "grad_norm": 0.06314355880022049,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 1.923574447631836,
+ "step": 2145
+ },
+ {
+ "epoch": 29.808362369337978,
+ "grad_norm": 0.0639389157295227,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 1.9339193105697632,
+ "step": 2146
+ },
+ {
+ "epoch": 29.822299651567945,
+ "grad_norm": 0.06419187039136887,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 1.9294497966766357,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83623693379791,
+ "grad_norm": 0.06351246684789658,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 1.9282407760620117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.850174216027874,
+ "grad_norm": 0.06371650099754333,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 1.914009690284729,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86411149825784,
+ "grad_norm": 0.06891132146120071,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 1.948460578918457,
+ "step": 2150
+ },
+ {
+ "epoch": 29.878048780487806,
+ "grad_norm": 0.06575452536344528,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 1.9238896369934082,
+ "step": 2151
+ },
+ {
+ "epoch": 29.89198606271777,
+ "grad_norm": 0.06279128044843674,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 1.9414188861846924,
+ "step": 2152
+ },
+ {
+ "epoch": 29.905923344947734,
+ "grad_norm": 0.06493876129388809,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 1.9337334632873535,
+ "step": 2153
+ },
+ {
+ "epoch": 29.9198606271777,
+ "grad_norm": 0.0640307068824768,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 1.9258478879928589,
+ "step": 2154
+ },
+ {
+ "epoch": 29.933797909407666,
+ "grad_norm": 0.06517527252435684,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 1.9236748218536377,
+ "step": 2155
+ },
+ {
+ "epoch": 29.94773519163763,
+ "grad_norm": 0.06373083591461182,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 1.9360520839691162,
+ "step": 2156
+ },
+ {
+ "epoch": 29.961672473867594,
+ "grad_norm": 0.06771238893270493,
+ "learning_rate": 0.00022244633283095,
+ "loss": 1.947780966758728,
+ "step": 2157
+ },
+ {
+ "epoch": 29.975609756097562,
+ "grad_norm": 0.06298559159040451,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 1.9223754405975342,
+ "step": 2158
+ },
+ {
+ "epoch": 29.989547038327526,
+ "grad_norm": 0.06455206871032715,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 1.9330079555511475,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.0522056445479393,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 1.4463932514190674,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6358165740966797,
+ "eval_runtime": 40.2782,
+ "eval_samples_per_second": 60.628,
+ "eval_steps_per_second": 0.497,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013937282229964,
+ "grad_norm": 0.06026934087276459,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 1.9035022258758545,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027874564459932,
+ "grad_norm": 0.06600011885166168,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 1.8965950012207031,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041811846689896,
+ "grad_norm": 0.060655586421489716,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 1.8954249620437622,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05574912891986,
+ "grad_norm": 0.0636422410607338,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 1.8951303958892822,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069686411149824,
+ "grad_norm": 0.05972317233681679,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 1.910957932472229,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083623693379792,
+ "grad_norm": 0.060592107474803925,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 1.9128470420837402,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097560975609756,
+ "grad_norm": 0.05888872966170311,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 1.897517204284668,
+ "step": 2167
+ },
+ {
+ "epoch": 30.11149825783972,
+ "grad_norm": 0.062287602573633194,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 1.9048705101013184,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125435540069688,
+ "grad_norm": 0.05919402465224266,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 1.901914119720459,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139372822299652,
+ "grad_norm": 0.06069913133978844,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 1.9066851139068604,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153310104529616,
+ "grad_norm": 0.06046717241406441,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 1.8969058990478516,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16724738675958,
+ "grad_norm": 0.059624481946229935,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 1.8908498287200928,
+ "step": 2172
+ },
+ {
+ "epoch": 30.181184668989548,
+ "grad_norm": 0.06100156158208847,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 1.909839391708374,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195121951219512,
+ "grad_norm": 0.05801135674118996,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 1.9057464599609375,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209059233449477,
+ "grad_norm": 0.06198010593652725,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 1.9051663875579834,
+ "step": 2175
+ },
+ {
+ "epoch": 30.222996515679444,
+ "grad_norm": 0.06107710674405098,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 1.904172658920288,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23693379790941,
+ "grad_norm": 0.060085635632276535,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 1.8970673084259033,
+ "step": 2177
+ },
+ {
+ "epoch": 30.250871080139373,
+ "grad_norm": 0.05755467712879181,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 1.9092798233032227,
+ "step": 2178
+ },
+ {
+ "epoch": 30.264808362369337,
+ "grad_norm": 0.05827959626913071,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 1.9012596607208252,
+ "step": 2179
+ },
+ {
+ "epoch": 30.278745644599304,
+ "grad_norm": 0.0570731982588768,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 1.9027369022369385,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29268292682927,
+ "grad_norm": 0.05898377671837807,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 1.9284251928329468,
+ "step": 2181
+ },
+ {
+ "epoch": 30.306620209059233,
+ "grad_norm": 0.06119367107748985,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 1.8992201089859009,
+ "step": 2182
+ },
+ {
+ "epoch": 30.320557491289197,
+ "grad_norm": 0.05790785327553749,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 1.9106863737106323,
+ "step": 2183
+ },
+ {
+ "epoch": 30.334494773519165,
+ "grad_norm": 0.06101157143712044,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 1.9129860401153564,
+ "step": 2184
+ },
+ {
+ "epoch": 30.34843205574913,
+ "grad_norm": 0.062228068709373474,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 1.9075040817260742,
+ "step": 2185
+ },
+ {
+ "epoch": 30.362369337979093,
+ "grad_norm": 0.057724758982658386,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 1.9021648168563843,
+ "step": 2186
+ },
+ {
+ "epoch": 30.37630662020906,
+ "grad_norm": 0.05886731296777725,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 1.8955028057098389,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390243902439025,
+ "grad_norm": 0.0613555908203125,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 1.8902294635772705,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40418118466899,
+ "grad_norm": 0.059424031525850296,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 1.9234192371368408,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418118466898953,
+ "grad_norm": 0.059314578771591187,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 1.9120237827301025,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43205574912892,
+ "grad_norm": 0.059916652739048004,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 1.9118289947509766,
+ "step": 2191
+ },
+ {
+ "epoch": 30.445993031358885,
+ "grad_norm": 0.05817565694451332,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 1.9012054204940796,
+ "step": 2192
+ },
+ {
+ "epoch": 30.45993031358885,
+ "grad_norm": 0.061957065016031265,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 1.9110759496688843,
+ "step": 2193
+ },
+ {
+ "epoch": 30.473867595818817,
+ "grad_norm": 0.05899856239557266,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 1.9285509586334229,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48780487804878,
+ "grad_norm": 0.06001519039273262,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 1.9093801975250244,
+ "step": 2195
+ },
+ {
+ "epoch": 30.501742160278745,
+ "grad_norm": 0.060825154185295105,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 1.892765760421753,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51567944250871,
+ "grad_norm": 0.06136799231171608,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 1.9276010990142822,
+ "step": 2197
+ },
+ {
+ "epoch": 30.529616724738677,
+ "grad_norm": 0.05773079767823219,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 1.9083237648010254,
+ "step": 2198
+ },
+ {
+ "epoch": 30.54355400696864,
+ "grad_norm": 0.06120939180254936,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 1.8891932964324951,
+ "step": 2199
+ },
+ {
+ "epoch": 30.557491289198605,
+ "grad_norm": 0.06097486987709999,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 1.9084399938583374,
+ "step": 2200
+ },
+ {
+ "epoch": 30.571428571428573,
+ "grad_norm": 0.05973029136657715,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 1.918110966682434,
+ "step": 2201
+ },
+ {
+ "epoch": 30.585365853658537,
+ "grad_norm": 0.06335631012916565,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 1.9174532890319824,
+ "step": 2202
+ },
+ {
+ "epoch": 30.5993031358885,
+ "grad_norm": 0.05974910035729408,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 1.9107253551483154,
+ "step": 2203
+ },
+ {
+ "epoch": 30.613240418118465,
+ "grad_norm": 0.06281539052724838,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 1.9007208347320557,
+ "step": 2204
+ },
+ {
+ "epoch": 30.627177700348433,
+ "grad_norm": 0.062120143324136734,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 1.911742925643921,
+ "step": 2205
+ },
+ {
+ "epoch": 30.641114982578397,
+ "grad_norm": 0.06305570155382156,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 1.9121983051300049,
+ "step": 2206
+ },
+ {
+ "epoch": 30.65505226480836,
+ "grad_norm": 0.06112781539559364,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 1.908568263053894,
+ "step": 2207
+ },
+ {
+ "epoch": 30.66898954703833,
+ "grad_norm": 0.06050721928477287,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 1.9018139839172363,
+ "step": 2208
+ },
+ {
+ "epoch": 30.682926829268293,
+ "grad_norm": 0.05980826914310455,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 1.9104691743850708,
+ "step": 2209
+ },
+ {
+ "epoch": 30.696864111498257,
+ "grad_norm": 0.06286003440618515,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 1.899999737739563,
+ "step": 2210
+ },
+ {
+ "epoch": 30.71080139372822,
+ "grad_norm": 0.06179053336381912,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 1.8993868827819824,
+ "step": 2211
+ },
+ {
+ "epoch": 30.72473867595819,
+ "grad_norm": 0.06006763502955437,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 1.9067524671554565,
+ "step": 2212
+ },
+ {
+ "epoch": 30.738675958188153,
+ "grad_norm": 0.06104302033782005,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 1.9074416160583496,
+ "step": 2213
+ },
+ {
+ "epoch": 30.752613240418118,
+ "grad_norm": 0.059952687472105026,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 1.9049818515777588,
+ "step": 2214
+ },
+ {
+ "epoch": 30.766550522648085,
+ "grad_norm": 0.06469111144542694,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 1.9107921123504639,
+ "step": 2215
+ },
+ {
+ "epoch": 30.78048780487805,
+ "grad_norm": 0.06027381867170334,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 1.920630693435669,
+ "step": 2216
+ },
+ {
+ "epoch": 30.794425087108014,
+ "grad_norm": 0.06066177412867546,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 1.9131195545196533,
+ "step": 2217
+ },
+ {
+ "epoch": 30.808362369337978,
+ "grad_norm": 0.06533747911453247,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 1.9172463417053223,
+ "step": 2218
+ },
+ {
+ "epoch": 30.822299651567945,
+ "grad_norm": 0.061867427080869675,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 1.9100360870361328,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83623693379791,
+ "grad_norm": 0.06394216418266296,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 1.9218776226043701,
+ "step": 2220
+ },
+ {
+ "epoch": 30.850174216027874,
+ "grad_norm": 0.061471469700336456,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 1.901368498802185,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86411149825784,
+ "grad_norm": 0.06380587071180344,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 1.9187582731246948,
+ "step": 2222
+ },
+ {
+ "epoch": 30.878048780487806,
+ "grad_norm": 0.06244281679391861,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 1.9082577228546143,
+ "step": 2223
+ },
+ {
+ "epoch": 30.89198606271777,
+ "grad_norm": 0.06435921788215637,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 1.924039363861084,
+ "step": 2224
+ },
+ {
+ "epoch": 30.905923344947734,
+ "grad_norm": 0.06331372261047363,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 1.9083380699157715,
+ "step": 2225
+ },
+ {
+ "epoch": 30.9198606271777,
+ "grad_norm": 0.06393173336982727,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 1.9132580757141113,
+ "step": 2226
+ },
+ {
+ "epoch": 30.933797909407666,
+ "grad_norm": 0.06500784307718277,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 1.9109617471694946,
+ "step": 2227
+ },
+ {
+ "epoch": 30.94773519163763,
+ "grad_norm": 0.06389841437339783,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 1.926948070526123,
+ "step": 2228
+ },
+ {
+ "epoch": 30.961672473867594,
+ "grad_norm": 0.06343282759189606,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 1.9304101467132568,
+ "step": 2229
+ },
+ {
+ "epoch": 30.975609756097562,
+ "grad_norm": 0.06003900617361069,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 1.9066693782806396,
+ "step": 2230
+ },
+ {
+ "epoch": 30.989547038327526,
+ "grad_norm": 0.06355264037847519,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 1.9183707237243652,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0492861270904541,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 1.4322078227996826,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6370227932929993,
+ "eval_runtime": 40.5546,
+ "eval_samples_per_second": 60.215,
+ "eval_steps_per_second": 0.493,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013937282229964,
+ "grad_norm": 0.057478148490190506,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 1.887700080871582,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027874564459932,
+ "grad_norm": 0.06220356002449989,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 1.8875237703323364,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041811846689896,
+ "grad_norm": 0.05686607584357262,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 1.897495150566101,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05574912891986,
+ "grad_norm": 0.06147320941090584,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 1.8913912773132324,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069686411149824,
+ "grad_norm": 0.06075664609670639,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 1.8823286294937134,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083623693379792,
+ "grad_norm": 0.058347299695014954,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 1.8890151977539062,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097560975609756,
+ "grad_norm": 0.06388939172029495,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 1.8857264518737793,
+ "step": 2239
+ },
+ {
+ "epoch": 31.11149825783972,
+ "grad_norm": 0.05901556834578514,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 1.8779339790344238,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125435540069688,
+ "grad_norm": 0.06094517558813095,
+ "learning_rate": 0.000201141724176723,
+ "loss": 1.8811697959899902,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139372822299652,
+ "grad_norm": 0.060666318982839584,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 1.8911728858947754,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153310104529616,
+ "grad_norm": 0.05608760938048363,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 1.8804845809936523,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16724738675958,
+ "grad_norm": 0.06151990965008736,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 1.8901598453521729,
+ "step": 2244
+ },
+ {
+ "epoch": 31.181184668989548,
+ "grad_norm": 0.05805948004126549,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 1.8834660053253174,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195121951219512,
+ "grad_norm": 0.05967888608574867,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 1.8807945251464844,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209059233449477,
+ "grad_norm": 0.06295159459114075,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 1.8946118354797363,
+ "step": 2247
+ },
+ {
+ "epoch": 31.222996515679444,
+ "grad_norm": 0.057971853762865067,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 1.889994502067566,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23693379790941,
+ "grad_norm": 0.05941799655556679,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 1.884669542312622,
+ "step": 2249
+ },
+ {
+ "epoch": 31.250871080139373,
+ "grad_norm": 0.06173485890030861,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 1.8727672100067139,
+ "step": 2250
+ },
+ {
+ "epoch": 31.264808362369337,
+ "grad_norm": 0.05898866057395935,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 1.8795123100280762,
+ "step": 2251
+ },
+ {
+ "epoch": 31.278745644599304,
+ "grad_norm": 0.06072458624839783,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 1.8846805095672607,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29268292682927,
+ "grad_norm": 0.060174115002155304,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 1.8993717432022095,
+ "step": 2253
+ },
+ {
+ "epoch": 31.306620209059233,
+ "grad_norm": 0.061605289578437805,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 1.8877589702606201,
+ "step": 2254
+ },
+ {
+ "epoch": 31.320557491289197,
+ "grad_norm": 0.06246590241789818,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 1.8974242210388184,
+ "step": 2255
+ },
+ {
+ "epoch": 31.334494773519165,
+ "grad_norm": 0.060071494430303574,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 1.881445288658142,
+ "step": 2256
+ },
+ {
+ "epoch": 31.34843205574913,
+ "grad_norm": 0.060360442847013474,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 1.8938015699386597,
+ "step": 2257
+ },
+ {
+ "epoch": 31.362369337979093,
+ "grad_norm": 0.06192562356591225,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 1.8961126804351807,
+ "step": 2258
+ },
+ {
+ "epoch": 31.37630662020906,
+ "grad_norm": 0.05872832611203194,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 1.8974393606185913,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390243902439025,
+ "grad_norm": 0.05980943515896797,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 1.8886656761169434,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40418118466899,
+ "grad_norm": 0.0593055821955204,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 1.8888055086135864,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418118466898953,
+ "grad_norm": 0.059456177055835724,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 1.9045109748840332,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43205574912892,
+ "grad_norm": 0.06260113418102264,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 1.9062271118164062,
+ "step": 2263
+ },
+ {
+ "epoch": 31.445993031358885,
+ "grad_norm": 0.05866454914212227,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 1.8920798301696777,
+ "step": 2264
+ },
+ {
+ "epoch": 31.45993031358885,
+ "grad_norm": 0.060281749814748764,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 1.9048649072647095,
+ "step": 2265
+ },
+ {
+ "epoch": 31.473867595818817,
+ "grad_norm": 0.060433730483055115,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 1.8905510902404785,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48780487804878,
+ "grad_norm": 0.06074376404285431,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 1.890744686126709,
+ "step": 2267
+ },
+ {
+ "epoch": 31.501742160278745,
+ "grad_norm": 0.060131702572107315,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 1.893791913986206,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51567944250871,
+ "grad_norm": 0.05804680287837982,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 1.8838505744934082,
+ "step": 2269
+ },
+ {
+ "epoch": 31.529616724738677,
+ "grad_norm": 0.05933694168925285,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 1.897843360900879,
+ "step": 2270
+ },
+ {
+ "epoch": 31.54355400696864,
+ "grad_norm": 0.058537356555461884,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 1.8826205730438232,
+ "step": 2271
+ },
+ {
+ "epoch": 31.557491289198605,
+ "grad_norm": 0.060081496834754944,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 1.889431357383728,
+ "step": 2272
+ },
+ {
+ "epoch": 31.571428571428573,
+ "grad_norm": 0.05904357507824898,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 1.8965020179748535,
+ "step": 2273
+ },
+ {
+ "epoch": 31.585365853658537,
+ "grad_norm": 0.05924354866147041,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 1.894204020500183,
+ "step": 2274
+ },
+ {
+ "epoch": 31.5993031358885,
+ "grad_norm": 0.059307269752025604,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 1.890209436416626,
+ "step": 2275
+ },
+ {
+ "epoch": 31.613240418118465,
+ "grad_norm": 0.05969221144914627,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 1.8865056037902832,
+ "step": 2276
+ },
+ {
+ "epoch": 31.627177700348433,
+ "grad_norm": 0.060738109052181244,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 1.8825939893722534,
+ "step": 2277
+ },
+ {
+ "epoch": 31.641114982578397,
+ "grad_norm": 0.06001656875014305,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 1.8988475799560547,
+ "step": 2278
+ },
+ {
+ "epoch": 31.65505226480836,
+ "grad_norm": 0.061478059738874435,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 1.9071214199066162,
+ "step": 2279
+ },
+ {
+ "epoch": 31.66898954703833,
+ "grad_norm": 0.06089174002408981,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 1.8998515605926514,
+ "step": 2280
+ },
+ {
+ "epoch": 31.682926829268293,
+ "grad_norm": 0.05940974876284599,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 1.8907361030578613,
+ "step": 2281
+ },
+ {
+ "epoch": 31.696864111498257,
+ "grad_norm": 0.06015758961439133,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 1.9010521173477173,
+ "step": 2282
+ },
+ {
+ "epoch": 31.71080139372822,
+ "grad_norm": 0.06062757968902588,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 1.899493932723999,
+ "step": 2283
+ },
+ {
+ "epoch": 31.72473867595819,
+ "grad_norm": 0.060489553958177567,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 1.9095027446746826,
+ "step": 2284
+ },
+ {
+ "epoch": 31.738675958188153,
+ "grad_norm": 0.05949694290757179,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 1.900972604751587,
+ "step": 2285
+ },
+ {
+ "epoch": 31.752613240418118,
+ "grad_norm": 0.06269118189811707,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 1.891621708869934,
+ "step": 2286
+ },
+ {
+ "epoch": 31.766550522648085,
+ "grad_norm": 0.06033318489789963,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 1.8954113721847534,
+ "step": 2287
+ },
+ {
+ "epoch": 31.78048780487805,
+ "grad_norm": 0.0604838989675045,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 1.9047741889953613,
+ "step": 2288
+ },
+ {
+ "epoch": 31.794425087108014,
+ "grad_norm": 0.060087885707616806,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 1.8912427425384521,
+ "step": 2289
+ },
+ {
+ "epoch": 31.808362369337978,
+ "grad_norm": 0.059592097997665405,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 1.8874431848526,
+ "step": 2290
+ },
+ {
+ "epoch": 31.822299651567945,
+ "grad_norm": 0.06219378113746643,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 1.8843477964401245,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83623693379791,
+ "grad_norm": 0.06010756641626358,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 1.892686367034912,
+ "step": 2292
+ },
+ {
+ "epoch": 31.850174216027874,
+ "grad_norm": 0.061404358595609665,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 1.907541275024414,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86411149825784,
+ "grad_norm": 0.060503825545310974,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 1.891597032546997,
+ "step": 2294
+ },
+ {
+ "epoch": 31.878048780487806,
+ "grad_norm": 0.06197115406394005,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 1.8968595266342163,
+ "step": 2295
+ },
+ {
+ "epoch": 31.89198606271777,
+ "grad_norm": 0.06244310364127159,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 1.903473138809204,
+ "step": 2296
+ },
+ {
+ "epoch": 31.905923344947734,
+ "grad_norm": 0.060648202896118164,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 1.9044740200042725,
+ "step": 2297
+ },
+ {
+ "epoch": 31.9198606271777,
+ "grad_norm": 0.061299171298742294,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 1.8890066146850586,
+ "step": 2298
+ },
+ {
+ "epoch": 31.933797909407666,
+ "grad_norm": 0.06204016134142876,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 1.912150502204895,
+ "step": 2299
+ },
+ {
+ "epoch": 31.94773519163763,
+ "grad_norm": 0.06292745471000671,
+ "learning_rate": 0.000186516746349841,
+ "loss": 1.9133987426757812,
+ "step": 2300
+ },
+ {
+ "epoch": 31.961672473867594,
+ "grad_norm": 0.0633154958486557,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 1.9106671810150146,
+ "step": 2301
+ },
+ {
+ "epoch": 31.975609756097562,
+ "grad_norm": 0.06269625574350357,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 1.9090297222137451,
+ "step": 2302
+ },
+ {
+ "epoch": 31.989547038327526,
+ "grad_norm": 0.06060625612735748,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 1.8874908685684204,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.05050232633948326,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 1.4178190231323242,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6386013627052307,
+ "eval_runtime": 40.9718,
+ "eval_samples_per_second": 59.602,
+ "eval_steps_per_second": 0.488,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01393728222997,
+ "grad_norm": 0.05701690912246704,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 1.8772780895233154,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02787456445993,
+ "grad_norm": 0.05597634240984917,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 1.8772567510604858,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041811846689896,
+ "grad_norm": 0.05631265789270401,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 1.8660085201263428,
+ "step": 2307
+ },
+ {
+ "epoch": 32.055749128919864,
+ "grad_norm": 0.05604572966694832,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 1.876168966293335,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069686411149824,
+ "grad_norm": 0.056123036891222,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 1.8643467426300049,
+ "step": 2309
+ },
+ {
+ "epoch": 32.08362369337979,
+ "grad_norm": 0.05791968107223511,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 1.866549015045166,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09756097560975,
+ "grad_norm": 0.05521981045603752,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 1.8852781057357788,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11149825783972,
+ "grad_norm": 0.055954866111278534,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 1.8603641986846924,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12543554006969,
+ "grad_norm": 0.05931234359741211,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 1.8619375228881836,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13937282229965,
+ "grad_norm": 0.05743015930056572,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 1.8824023008346558,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153310104529616,
+ "grad_norm": 0.056393202394247055,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 1.8722079992294312,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167247386759584,
+ "grad_norm": 0.05671187862753868,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 1.8824248313903809,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181184668989545,
+ "grad_norm": 0.05667170509696007,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 1.882739782333374,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19512195121951,
+ "grad_norm": 0.05515167862176895,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 1.8555771112442017,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20905923344948,
+ "grad_norm": 0.05649147555232048,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 1.8673419952392578,
+ "step": 2319
+ },
+ {
+ "epoch": 32.22299651567944,
+ "grad_norm": 0.058623913675546646,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 1.8744664192199707,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23693379790941,
+ "grad_norm": 0.055131133645772934,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 1.8781135082244873,
+ "step": 2321
+ },
+ {
+ "epoch": 32.250871080139376,
+ "grad_norm": 0.06008831411600113,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 1.8844444751739502,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26480836236934,
+ "grad_norm": 0.05755964294075966,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 1.8790829181671143,
+ "step": 2323
+ },
+ {
+ "epoch": 32.278745644599304,
+ "grad_norm": 0.055803194642066956,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 1.8642692565917969,
+ "step": 2324
+ },
+ {
+ "epoch": 32.292682926829265,
+ "grad_norm": 0.05710228905081749,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 1.8659579753875732,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30662020905923,
+ "grad_norm": 0.05771065130829811,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 1.8812355995178223,
+ "step": 2326
+ },
+ {
+ "epoch": 32.3205574912892,
+ "grad_norm": 0.05642757564783096,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 1.8610783815383911,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33449477351916,
+ "grad_norm": 0.05687689036130905,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 1.875876545906067,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34843205574913,
+ "grad_norm": 0.057835984975099564,
+ "learning_rate": 0.000179445406945268,
+ "loss": 1.8674073219299316,
+ "step": 2329
+ },
+ {
+ "epoch": 32.362369337979096,
+ "grad_norm": 0.05800129100680351,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 1.8708086013793945,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37630662020906,
+ "grad_norm": 0.05744084715843201,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 1.8869245052337646,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390243902439025,
+ "grad_norm": 0.056703388690948486,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 1.8688050508499146,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40418118466899,
+ "grad_norm": 0.06032602861523628,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 1.873382568359375,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41811846689895,
+ "grad_norm": 0.05810776352882385,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 1.8624725341796875,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43205574912892,
+ "grad_norm": 0.060990724712610245,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 1.8792011737823486,
+ "step": 2335
+ },
+ {
+ "epoch": 32.44599303135889,
+ "grad_norm": 0.05975554883480072,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 1.8793275356292725,
+ "step": 2336
+ },
+ {
+ "epoch": 32.45993031358885,
+ "grad_norm": 0.05772533640265465,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 1.889489769935608,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47386759581882,
+ "grad_norm": 0.05920211225748062,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 1.8647074699401855,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48780487804878,
+ "grad_norm": 0.059745173901319504,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 1.8649309873580933,
+ "step": 2339
+ },
+ {
+ "epoch": 32.501742160278745,
+ "grad_norm": 0.05933282524347305,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 1.8936223983764648,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51567944250871,
+ "grad_norm": 0.062407393008470535,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 1.8831725120544434,
+ "step": 2341
+ },
+ {
+ "epoch": 32.52961672473867,
+ "grad_norm": 0.061185337603092194,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 1.8766106367111206,
+ "step": 2342
+ },
+ {
+ "epoch": 32.54355400696864,
+ "grad_norm": 0.06082421913743019,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 1.880276083946228,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55749128919861,
+ "grad_norm": 0.05900120362639427,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 1.8696876764297485,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57142857142857,
+ "grad_norm": 0.06146271899342537,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 1.864588737487793,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58536585365854,
+ "grad_norm": 0.05976773053407669,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 1.8771488666534424,
+ "step": 2346
+ },
+ {
+ "epoch": 32.599303135888505,
+ "grad_norm": 0.05808904021978378,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 1.890838623046875,
+ "step": 2347
+ },
+ {
+ "epoch": 32.613240418118465,
+ "grad_norm": 0.06149672716856003,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 1.8787143230438232,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62717770034843,
+ "grad_norm": 0.059028271585702896,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 1.8875718116760254,
+ "step": 2349
+ },
+ {
+ "epoch": 32.641114982578394,
+ "grad_norm": 0.062163304537534714,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 1.86748206615448,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65505226480836,
+ "grad_norm": 0.06221195310354233,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 1.8742306232452393,
+ "step": 2351
+ },
+ {
+ "epoch": 32.66898954703833,
+ "grad_norm": 0.060461003333330154,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 1.885633945465088,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68292682926829,
+ "grad_norm": 0.06130106747150421,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 1.8827314376831055,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69686411149826,
+ "grad_norm": 0.06044832617044449,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 1.8724548816680908,
+ "step": 2354
+ },
+ {
+ "epoch": 32.710801393728225,
+ "grad_norm": 0.06029713526368141,
+ "learning_rate": 0.000173176617304673,
+ "loss": 1.8745346069335938,
+ "step": 2355
+ },
+ {
+ "epoch": 32.724738675958186,
+ "grad_norm": 0.06287439167499542,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 1.8879387378692627,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73867595818815,
+ "grad_norm": 0.059636496007442474,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 1.882830262184143,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75261324041812,
+ "grad_norm": 0.06590992957353592,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 1.882242202758789,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76655052264808,
+ "grad_norm": 0.06283026933670044,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 1.8864011764526367,
+ "step": 2359
+ },
+ {
+ "epoch": 32.78048780487805,
+ "grad_norm": 0.061717696487903595,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 1.8815748691558838,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79442508710802,
+ "grad_norm": 0.06423293799161911,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 1.8824026584625244,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80836236933798,
+ "grad_norm": 0.06351540237665176,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 1.9000264406204224,
+ "step": 2362
+ },
+ {
+ "epoch": 32.822299651567945,
+ "grad_norm": 0.060434527695178986,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 1.8875732421875,
+ "step": 2363
+ },
+ {
+ "epoch": 32.836236933797906,
+ "grad_norm": 0.062305059283971786,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 1.8695244789123535,
+ "step": 2364
+ },
+ {
+ "epoch": 32.850174216027874,
+ "grad_norm": 0.05981886014342308,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 1.8885269165039062,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86411149825784,
+ "grad_norm": 0.0642906129360199,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 1.8818528652191162,
+ "step": 2366
+ },
+ {
+ "epoch": 32.8780487804878,
+ "grad_norm": 0.06316667050123215,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 1.8994319438934326,
+ "step": 2367
+ },
+ {
+ "epoch": 32.89198606271777,
+ "grad_norm": 0.058260831981897354,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 1.8908791542053223,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90592334494774,
+ "grad_norm": 0.062237389385700226,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 1.8753025531768799,
+ "step": 2369
+ },
+ {
+ "epoch": 32.9198606271777,
+ "grad_norm": 0.05972493812441826,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 1.899425745010376,
+ "step": 2370
+ },
+ {
+ "epoch": 32.933797909407666,
+ "grad_norm": 0.058846332132816315,
+ "learning_rate": 0.00016935382741164,
+ "loss": 1.8934335708618164,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94773519163763,
+ "grad_norm": 0.06333424896001816,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 1.8951395750045776,
+ "step": 2372
+ },
+ {
+ "epoch": 32.961672473867594,
+ "grad_norm": 0.06063693389296532,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 1.881688117980957,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97560975609756,
+ "grad_norm": 0.060873813927173615,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 1.8836770057678223,
+ "step": 2374
+ },
+ {
+ "epoch": 32.98954703832753,
+ "grad_norm": 0.060378748923540115,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 1.886624813079834,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.04906065762042999,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 1.4146361351013184,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6401442885398865,
+ "eval_runtime": 41.0841,
+ "eval_samples_per_second": 59.439,
+ "eval_steps_per_second": 0.487,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01393728222997,
+ "grad_norm": 0.05557359382510185,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 1.8605964183807373,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02787456445993,
+ "grad_norm": 0.055585991591215134,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 1.8456363677978516,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041811846689896,
+ "grad_norm": 0.05610295757651329,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 1.847888469696045,
+ "step": 2379
+ },
+ {
+ "epoch": 33.055749128919864,
+ "grad_norm": 0.05532662197947502,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 1.850597620010376,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069686411149824,
+ "grad_norm": 0.05585319921374321,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 1.8694689273834229,
+ "step": 2381
+ },
+ {
+ "epoch": 33.08362369337979,
+ "grad_norm": 0.05658521130681038,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 1.8533908128738403,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09756097560975,
+ "grad_norm": 0.0558413490653038,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 1.8735582828521729,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11149825783972,
+ "grad_norm": 0.05637028068304062,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 1.8585509061813354,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12543554006969,
+ "grad_norm": 0.055854618549346924,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 1.8539409637451172,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13937282229965,
+ "grad_norm": 0.055990077555179596,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 1.85886812210083,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153310104529616,
+ "grad_norm": 0.05781752988696098,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 1.8603525161743164,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167247386759584,
+ "grad_norm": 0.055855557322502136,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 1.861933946609497,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181184668989545,
+ "grad_norm": 0.05729927122592926,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 1.86094069480896,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19512195121951,
+ "grad_norm": 0.056495532393455505,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 1.848013997077942,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20905923344948,
+ "grad_norm": 0.0571073479950428,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 1.8558173179626465,
+ "step": 2391
+ },
+ {
+ "epoch": 33.22299651567944,
+ "grad_norm": 0.056066226214170456,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 1.8723382949829102,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23693379790941,
+ "grad_norm": 0.05570382624864578,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 1.864349365234375,
+ "step": 2393
+ },
+ {
+ "epoch": 33.250871080139376,
+ "grad_norm": 0.054912492632865906,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 1.8678479194641113,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26480836236934,
+ "grad_norm": 0.05622183158993721,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 1.8648667335510254,
+ "step": 2395
+ },
+ {
+ "epoch": 33.278745644599304,
+ "grad_norm": 0.05719740688800812,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 1.850623607635498,
+ "step": 2396
+ },
+ {
+ "epoch": 33.292682926829265,
+ "grad_norm": 0.056594349443912506,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 1.8707754611968994,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30662020905923,
+ "grad_norm": 0.05596613511443138,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 1.8580527305603027,
+ "step": 2398
+ },
+ {
+ "epoch": 33.3205574912892,
+ "grad_norm": 0.05675627663731575,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 1.8687434196472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33449477351916,
+ "grad_norm": 0.05572087690234184,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 1.8603391647338867,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34843205574913,
+ "grad_norm": 0.05837463587522507,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 1.8638341426849365,
+ "step": 2401
+ },
+ {
+ "epoch": 33.362369337979096,
+ "grad_norm": 0.056136030703783035,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 1.853350281715393,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37630662020906,
+ "grad_norm": 0.0555158406496048,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 1.8775548934936523,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390243902439025,
+ "grad_norm": 0.0595950111746788,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 1.8578253984451294,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40418118466899,
+ "grad_norm": 0.05565760284662247,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 1.8508214950561523,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41811846689895,
+ "grad_norm": 0.058832235634326935,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 1.8614635467529297,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43205574912892,
+ "grad_norm": 0.0585586316883564,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 1.8617370128631592,
+ "step": 2407
+ },
+ {
+ "epoch": 33.44599303135889,
+ "grad_norm": 0.05951899662613869,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 1.8648651838302612,
+ "step": 2408
+ },
+ {
+ "epoch": 33.45993031358885,
+ "grad_norm": 0.057204652577638626,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 1.8577271699905396,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47386759581882,
+ "grad_norm": 0.05844349041581154,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 1.8560724258422852,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48780487804878,
+ "grad_norm": 0.05761651322245598,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 1.863309383392334,
+ "step": 2411
+ },
+ {
+ "epoch": 33.501742160278745,
+ "grad_norm": 0.05818328261375427,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 1.8798553943634033,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51567944250871,
+ "grad_norm": 0.05746658146381378,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 1.8628616333007812,
+ "step": 2413
+ },
+ {
+ "epoch": 33.52961672473867,
+ "grad_norm": 0.057656314224004745,
+ "learning_rate": 0.000159218843594243,
+ "loss": 1.8699405193328857,
+ "step": 2414
+ },
+ {
+ "epoch": 33.54355400696864,
+ "grad_norm": 0.0586664155125618,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 1.8481391668319702,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55749128919861,
+ "grad_norm": 0.05769997090101242,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 1.8612866401672363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57142857142857,
+ "grad_norm": 0.0598456896841526,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 1.8580410480499268,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58536585365854,
+ "grad_norm": 0.05761442333459854,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 1.874891996383667,
+ "step": 2418
+ },
+ {
+ "epoch": 33.599303135888505,
+ "grad_norm": 0.06176742538809776,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 1.8660908937454224,
+ "step": 2419
+ },
+ {
+ "epoch": 33.613240418118465,
+ "grad_norm": 0.0588943213224411,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 1.8718886375427246,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62717770034843,
+ "grad_norm": 0.06092666834592819,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 1.8634107112884521,
+ "step": 2421
+ },
+ {
+ "epoch": 33.641114982578394,
+ "grad_norm": 0.05774684250354767,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 1.868361234664917,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65505226480836,
+ "grad_norm": 0.05932781100273132,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 1.867889642715454,
+ "step": 2423
+ },
+ {
+ "epoch": 33.66898954703833,
+ "grad_norm": 0.058603718876838684,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 1.8793346881866455,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68292682926829,
+ "grad_norm": 0.05870316922664642,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 1.8717231750488281,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69686411149826,
+ "grad_norm": 0.05805273354053497,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 1.8479297161102295,
+ "step": 2426
+ },
+ {
+ "epoch": 33.710801393728225,
+ "grad_norm": 0.059018444269895554,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 1.8711276054382324,
+ "step": 2427
+ },
+ {
+ "epoch": 33.724738675958186,
+ "grad_norm": 0.05927985534071922,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 1.8653428554534912,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73867595818815,
+ "grad_norm": 0.06085766851902008,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 1.8724851608276367,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75261324041812,
+ "grad_norm": 0.05974132940173149,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 1.8459011316299438,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76655052264808,
+ "grad_norm": 0.05851755291223526,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 1.863109827041626,
+ "step": 2431
+ },
+ {
+ "epoch": 33.78048780487805,
+ "grad_norm": 0.06079421192407608,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 1.8687033653259277,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79442508710802,
+ "grad_norm": 0.057930078357458115,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 1.8653396368026733,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80836236933798,
+ "grad_norm": 0.059440065175294876,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 1.8569486141204834,
+ "step": 2434
+ },
+ {
+ "epoch": 33.822299651567945,
+ "grad_norm": 0.05835104361176491,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 1.8828535079956055,
+ "step": 2435
+ },
+ {
+ "epoch": 33.836236933797906,
+ "grad_norm": 0.06108623370528221,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 1.8749947547912598,
+ "step": 2436
+ },
+ {
+ "epoch": 33.850174216027874,
+ "grad_norm": 0.059921979904174805,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 1.8684660196304321,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86411149825784,
+ "grad_norm": 0.059964295476675034,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 1.862084150314331,
+ "step": 2438
+ },
+ {
+ "epoch": 33.8780487804878,
+ "grad_norm": 0.05888533219695091,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 1.85898756980896,
+ "step": 2439
+ },
+ {
+ "epoch": 33.89198606271777,
+ "grad_norm": 0.058946020901203156,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 1.872513771057129,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90592334494774,
+ "grad_norm": 0.05947734788060188,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 1.8610495328903198,
+ "step": 2441
+ },
+ {
+ "epoch": 33.9198606271777,
+ "grad_norm": 0.05859588086605072,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 1.8842005729675293,
+ "step": 2442
+ },
+ {
+ "epoch": 33.933797909407666,
+ "grad_norm": 0.05818016827106476,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 1.8715083599090576,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94773519163763,
+ "grad_norm": 0.05960588902235031,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 1.881840467453003,
+ "step": 2444
+ },
+ {
+ "epoch": 33.961672473867594,
+ "grad_norm": 0.059536516666412354,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 1.8734736442565918,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97560975609756,
+ "grad_norm": 0.05922378972172737,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 1.8569964170455933,
+ "step": 2446
+ },
+ {
+ "epoch": 33.98954703832753,
+ "grad_norm": 0.058573752641677856,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 1.8710730075836182,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.047325484454631805,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 1.4022152423858643,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6414204835891724,
+ "eval_runtime": 39.9489,
+ "eval_samples_per_second": 61.128,
+ "eval_steps_per_second": 0.501,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01393728222997,
+ "grad_norm": 0.05358244478702545,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 1.8384610414505005,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02787456445993,
+ "grad_norm": 0.0558239221572876,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 1.8538508415222168,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041811846689896,
+ "grad_norm": 0.05390622094273567,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 1.8460195064544678,
+ "step": 2451
+ },
+ {
+ "epoch": 34.055749128919864,
+ "grad_norm": 0.056224383413791656,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 1.8416059017181396,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069686411149824,
+ "grad_norm": 0.053925771266222,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 1.8532803058624268,
+ "step": 2453
+ },
+ {
+ "epoch": 34.08362369337979,
+ "grad_norm": 0.05750057101249695,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 1.847919225692749,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09756097560975,
+ "grad_norm": 0.05499042198061943,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 1.8450297117233276,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11149825783972,
+ "grad_norm": 0.0544770173728466,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 1.8482413291931152,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12543554006969,
+ "grad_norm": 0.05656695365905762,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 1.860766053199768,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13937282229965,
+ "grad_norm": 0.054617058485746384,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 1.8398511409759521,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153310104529616,
+ "grad_norm": 0.05470099300146103,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 1.8501665592193604,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167247386759584,
+ "grad_norm": 0.05437156930565834,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 1.8514611721038818,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181184668989545,
+ "grad_norm": 0.05481892451643944,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 1.8424533605575562,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19512195121951,
+ "grad_norm": 0.055016402155160904,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 1.823396921157837,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20905923344948,
+ "grad_norm": 0.05608086287975311,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 1.8525217771530151,
+ "step": 2463
+ },
+ {
+ "epoch": 34.22299651567944,
+ "grad_norm": 0.055000413209199905,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 1.8570690155029297,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23693379790941,
+ "grad_norm": 0.05437581241130829,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 1.8471980094909668,
+ "step": 2465
+ },
+ {
+ "epoch": 34.250871080139376,
+ "grad_norm": 0.0553814135491848,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 1.8384971618652344,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26480836236934,
+ "grad_norm": 0.053538259118795395,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 1.8536994457244873,
+ "step": 2467
+ },
+ {
+ "epoch": 34.278745644599304,
+ "grad_norm": 0.05509314313530922,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 1.8386962413787842,
+ "step": 2468
+ },
+ {
+ "epoch": 34.292682926829265,
+ "grad_norm": 0.054552506655454636,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 1.8502833843231201,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30662020905923,
+ "grad_norm": 0.054326511919498444,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 1.8350508213043213,
+ "step": 2470
+ },
+ {
+ "epoch": 34.3205574912892,
+ "grad_norm": 0.05458610877394676,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 1.8488143682479858,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33449477351916,
+ "grad_norm": 0.05537666752934456,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 1.8561567068099976,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34843205574913,
+ "grad_norm": 0.056333623826503754,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 1.8610734939575195,
+ "step": 2473
+ },
+ {
+ "epoch": 34.362369337979096,
+ "grad_norm": 0.053708501160144806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 1.8342688083648682,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37630662020906,
+ "grad_norm": 0.05496754124760628,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 1.8592373132705688,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390243902439025,
+ "grad_norm": 0.05445944517850876,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 1.8551034927368164,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40418118466899,
+ "grad_norm": 0.055870212614536285,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 1.8510769605636597,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41811846689895,
+ "grad_norm": 0.05536561831831932,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 1.8463027477264404,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43205574912892,
+ "grad_norm": 0.054308436810970306,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 1.8470419645309448,
+ "step": 2479
+ },
+ {
+ "epoch": 34.44599303135889,
+ "grad_norm": 0.056683722883462906,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 1.8605597019195557,
+ "step": 2480
+ },
+ {
+ "epoch": 34.45993031358885,
+ "grad_norm": 0.05601316690444946,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 1.847717046737671,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47386759581882,
+ "grad_norm": 0.05449012294411659,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 1.8464031219482422,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48780487804878,
+ "grad_norm": 0.054641272872686386,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 1.832083821296692,
+ "step": 2483
+ },
+ {
+ "epoch": 34.501742160278745,
+ "grad_norm": 0.05610528588294983,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 1.8700275421142578,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51567944250871,
+ "grad_norm": 0.0551300011575222,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 1.8514257669448853,
+ "step": 2485
+ },
+ {
+ "epoch": 34.52961672473867,
+ "grad_norm": 0.056593794375658035,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 1.8480379581451416,
+ "step": 2486
+ },
+ {
+ "epoch": 34.54355400696864,
+ "grad_norm": 0.055673424154520035,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 1.838927984237671,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55749128919861,
+ "grad_norm": 0.05493558943271637,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 1.8383840322494507,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57142857142857,
+ "grad_norm": 0.05680295452475548,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 1.8514118194580078,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58536585365854,
+ "grad_norm": 0.057038743048906326,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 1.8620257377624512,
+ "step": 2490
+ },
+ {
+ "epoch": 34.599303135888505,
+ "grad_norm": 0.05735227093100548,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 1.8413968086242676,
+ "step": 2491
+ },
+ {
+ "epoch": 34.613240418118465,
+ "grad_norm": 0.058245740830898285,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 1.8599750995635986,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62717770034843,
+ "grad_norm": 0.056244585663080215,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 1.8527486324310303,
+ "step": 2493
+ },
+ {
+ "epoch": 34.641114982578394,
+ "grad_norm": 0.05616258084774017,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 1.84233558177948,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65505226480836,
+ "grad_norm": 0.057303108274936676,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 1.865498661994934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.66898954703833,
+ "grad_norm": 0.05895329639315605,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 1.8521645069122314,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68292682926829,
+ "grad_norm": 0.05679212138056755,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 1.8539199829101562,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69686411149826,
+ "grad_norm": 0.05897561088204384,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 1.8534815311431885,
+ "step": 2498
+ },
+ {
+ "epoch": 34.710801393728225,
+ "grad_norm": 0.05891786143183708,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 1.845531702041626,
+ "step": 2499
+ },
+ {
+ "epoch": 34.724738675958186,
+ "grad_norm": 0.059177134186029434,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 1.8502740859985352,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73867595818815,
+ "grad_norm": 0.05862346291542053,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 1.8657329082489014,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75261324041812,
+ "grad_norm": 0.061109986156225204,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 1.8452383279800415,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76655052264808,
+ "grad_norm": 0.058431629091501236,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 1.8575074672698975,
+ "step": 2503
+ },
+ {
+ "epoch": 34.78048780487805,
+ "grad_norm": 0.05970519408583641,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 1.8476191759109497,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79442508710802,
+ "grad_norm": 0.05823173746466637,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 1.8554785251617432,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80836236933798,
+ "grad_norm": 0.060084082186222076,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 1.8471784591674805,
+ "step": 2506
+ },
+ {
+ "epoch": 34.822299651567945,
+ "grad_norm": 0.0589512400329113,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 1.86192786693573,
+ "step": 2507
+ },
+ {
+ "epoch": 34.836236933797906,
+ "grad_norm": 0.06184513866901398,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 1.8572009801864624,
+ "step": 2508
+ },
+ {
+ "epoch": 34.850174216027874,
+ "grad_norm": 0.05833953991532326,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 1.8561553955078125,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86411149825784,
+ "grad_norm": 0.061172984540462494,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 1.8583234548568726,
+ "step": 2510
+ },
+ {
+ "epoch": 34.8780487804878,
+ "grad_norm": 0.0573224313557148,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 1.8403494358062744,
+ "step": 2511
+ },
+ {
+ "epoch": 34.89198606271777,
+ "grad_norm": 0.06135294586420059,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 1.869114875793457,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90592334494774,
+ "grad_norm": 0.05907885357737541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 1.8581886291503906,
+ "step": 2513
+ },
+ {
+ "epoch": 34.9198606271777,
+ "grad_norm": 0.06035003438591957,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 1.8454012870788574,
+ "step": 2514
+ },
+ {
+ "epoch": 34.933797909407666,
+ "grad_norm": 0.060585640370845795,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 1.8453309535980225,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94773519163763,
+ "grad_norm": 0.06050759181380272,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 1.8467144966125488,
+ "step": 2516
+ },
+ {
+ "epoch": 34.961672473867594,
+ "grad_norm": 0.06118239462375641,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 1.8623638153076172,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97560975609756,
+ "grad_norm": 0.059003446251153946,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 1.8540363311767578,
+ "step": 2518
+ },
+ {
+ "epoch": 34.98954703832753,
+ "grad_norm": 0.06078164279460907,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 1.8434364795684814,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04768386855721474,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 1.3819507360458374,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.642964243888855,
+ "eval_runtime": 40.0846,
+ "eval_samples_per_second": 60.921,
+ "eval_steps_per_second": 0.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01393728222997,
+ "grad_norm": 0.05531901493668556,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 1.8454805612564087,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02787456445993,
+ "grad_norm": 0.05443203076720238,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 1.8266499042510986,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041811846689896,
+ "grad_norm": 0.05491992458701134,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 1.839545488357544,
+ "step": 2523
+ },
+ {
+ "epoch": 35.055749128919864,
+ "grad_norm": 0.05408782139420509,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 1.836404800415039,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069686411149824,
+ "grad_norm": 0.0541234128177166,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 1.8264034986495972,
+ "step": 2525
+ },
+ {
+ "epoch": 35.08362369337979,
+ "grad_norm": 0.05504459887742996,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 1.8485287427902222,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09756097560975,
+ "grad_norm": 0.05482470244169235,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 1.8336119651794434,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11149825783972,
+ "grad_norm": 0.05490747466683388,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 1.8223836421966553,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12543554006969,
+ "grad_norm": 0.05421694368124008,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 1.8309847116470337,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13937282229965,
+ "grad_norm": 0.05432746186852455,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 1.8214728832244873,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153310104529616,
+ "grad_norm": 0.05329402908682823,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 1.8356881141662598,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167247386759584,
+ "grad_norm": 0.05515717342495918,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 1.8358293771743774,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181184668989545,
+ "grad_norm": 0.05425635725259781,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 1.8230032920837402,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19512195121951,
+ "grad_norm": 0.054918598383665085,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 1.8427648544311523,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20905923344948,
+ "grad_norm": 0.05530979484319687,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 1.811009407043457,
+ "step": 2535
+ },
+ {
+ "epoch": 35.22299651567944,
+ "grad_norm": 0.05482922121882439,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 1.8356122970581055,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23693379790941,
+ "grad_norm": 0.05470896139740944,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 1.8432835340499878,
+ "step": 2537
+ },
+ {
+ "epoch": 35.250871080139376,
+ "grad_norm": 0.05382648482918739,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 1.8248438835144043,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26480836236934,
+ "grad_norm": 0.05452217534184456,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 1.8385814428329468,
+ "step": 2539
+ },
+ {
+ "epoch": 35.278745644599304,
+ "grad_norm": 0.05492600053548813,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 1.8321887254714966,
+ "step": 2540
+ },
+ {
+ "epoch": 35.292682926829265,
+ "grad_norm": 0.054316446185112,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 1.8411835432052612,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30662020905923,
+ "grad_norm": 0.05490174889564514,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 1.8350071907043457,
+ "step": 2542
+ },
+ {
+ "epoch": 35.3205574912892,
+ "grad_norm": 0.055120889097452164,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 1.8385683298110962,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33449477351916,
+ "grad_norm": 0.054340824484825134,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 1.856696605682373,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34843205574913,
+ "grad_norm": 0.056243546307086945,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 1.8386306762695312,
+ "step": 2545
+ },
+ {
+ "epoch": 35.362369337979096,
+ "grad_norm": 0.05486825481057167,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 1.8275195360183716,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37630662020906,
+ "grad_norm": 0.05750235170125961,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 1.8266764879226685,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390243902439025,
+ "grad_norm": 0.05539723485708237,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 1.8288836479187012,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40418118466899,
+ "grad_norm": 0.056604526937007904,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 1.8384013175964355,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41811846689895,
+ "grad_norm": 0.05410664901137352,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 1.8481371402740479,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43205574912892,
+ "grad_norm": 0.0576784573495388,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 1.8333299160003662,
+ "step": 2551
+ },
+ {
+ "epoch": 35.44599303135889,
+ "grad_norm": 0.05358795076608658,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 1.8201930522918701,
+ "step": 2552
+ },
+ {
+ "epoch": 35.45993031358885,
+ "grad_norm": 0.05687721073627472,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 1.838738203048706,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47386759581882,
+ "grad_norm": 0.05585115775465965,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 1.8358794450759888,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48780487804878,
+ "grad_norm": 0.057034578174352646,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 1.8355228900909424,
+ "step": 2555
+ },
+ {
+ "epoch": 35.501742160278745,
+ "grad_norm": 0.05627685412764549,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 1.8351898193359375,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51567944250871,
+ "grad_norm": 0.05724332481622696,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 1.833801031112671,
+ "step": 2557
+ },
+ {
+ "epoch": 35.52961672473867,
+ "grad_norm": 0.05503929778933525,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 1.8311514854431152,
+ "step": 2558
+ },
+ {
+ "epoch": 35.54355400696864,
+ "grad_norm": 0.057043809443712234,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 1.8318891525268555,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55749128919861,
+ "grad_norm": 0.05669008940458298,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 1.8333967924118042,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57142857142857,
+ "grad_norm": 0.05707597732543945,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 1.8416266441345215,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58536585365854,
+ "grad_norm": 0.055258579552173615,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 1.8428752422332764,
+ "step": 2562
+ },
+ {
+ "epoch": 35.599303135888505,
+ "grad_norm": 0.05575427785515785,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 1.8269420862197876,
+ "step": 2563
+ },
+ {
+ "epoch": 35.613240418118465,
+ "grad_norm": 0.05566711723804474,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 1.8391327857971191,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62717770034843,
+ "grad_norm": 0.057135093957185745,
+ "learning_rate": 0.000125422220031917,
+ "loss": 1.8499560356140137,
+ "step": 2565
+ },
+ {
+ "epoch": 35.641114982578394,
+ "grad_norm": 0.05612964928150177,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 1.8274946212768555,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65505226480836,
+ "grad_norm": 0.05745973810553551,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 1.818924903869629,
+ "step": 2567
+ },
+ {
+ "epoch": 35.66898954703833,
+ "grad_norm": 0.05643341690301895,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 1.8365602493286133,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68292682926829,
+ "grad_norm": 0.054355110973119736,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 1.8275706768035889,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69686411149826,
+ "grad_norm": 0.05827079713344574,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 1.8421577215194702,
+ "step": 2570
+ },
+ {
+ "epoch": 35.710801393728225,
+ "grad_norm": 0.055892039090394974,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 1.8322241306304932,
+ "step": 2571
+ },
+ {
+ "epoch": 35.724738675958186,
+ "grad_norm": 0.05828035622835159,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 1.8444736003875732,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73867595818815,
+ "grad_norm": 0.05669068545103073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 1.8492958545684814,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75261324041812,
+ "grad_norm": 0.05769318342208862,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 1.8362290859222412,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76655052264808,
+ "grad_norm": 0.059012528508901596,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 1.8452755212783813,
+ "step": 2575
+ },
+ {
+ "epoch": 35.78048780487805,
+ "grad_norm": 0.0577283576130867,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 1.8381261825561523,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79442508710802,
+ "grad_norm": 0.05785193666815758,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 1.8344440460205078,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80836236933798,
+ "grad_norm": 0.057778336107730865,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 1.8465049266815186,
+ "step": 2578
+ },
+ {
+ "epoch": 35.822299651567945,
+ "grad_norm": 0.05725527182221413,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 1.8357681035995483,
+ "step": 2579
+ },
+ {
+ "epoch": 35.836236933797906,
+ "grad_norm": 0.05706310644745827,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 1.8430027961730957,
+ "step": 2580
+ },
+ {
+ "epoch": 35.850174216027874,
+ "grad_norm": 0.05754264071583748,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 1.8457331657409668,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86411149825784,
+ "grad_norm": 0.05796574801206589,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 1.8521018028259277,
+ "step": 2582
+ },
+ {
+ "epoch": 35.8780487804878,
+ "grad_norm": 0.05767996609210968,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 1.837098240852356,
+ "step": 2583
+ },
+ {
+ "epoch": 35.89198606271777,
+ "grad_norm": 0.05741669982671738,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 1.8438975811004639,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90592334494774,
+ "grad_norm": 0.05797840282320976,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 1.8390898704528809,
+ "step": 2585
+ },
+ {
+ "epoch": 35.9198606271777,
+ "grad_norm": 0.057005491107702255,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 1.853393316268921,
+ "step": 2586
+ },
+ {
+ "epoch": 35.933797909407666,
+ "grad_norm": 0.058777451515197754,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 1.8450913429260254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94773519163763,
+ "grad_norm": 0.05810392275452614,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 1.8371872901916504,
+ "step": 2588
+ },
+ {
+ "epoch": 35.961672473867594,
+ "grad_norm": 0.05875588208436966,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 1.8611619472503662,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97560975609756,
+ "grad_norm": 0.05872174724936485,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 1.8622443675994873,
+ "step": 2590
+ },
+ {
+ "epoch": 35.98954703832753,
+ "grad_norm": 0.059304554015398026,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 1.8395988941192627,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04685443267226219,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 1.3844654560089111,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.644555926322937,
+ "eval_runtime": 40.5311,
+ "eval_samples_per_second": 60.25,
+ "eval_steps_per_second": 0.493,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01393728222997,
+ "grad_norm": 0.05542989447712898,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 1.816087007522583,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02787456445993,
+ "grad_norm": 0.05269062519073486,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 1.8240373134613037,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041811846689896,
+ "grad_norm": 0.05406419187784195,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 1.8276968002319336,
+ "step": 2595
+ },
+ {
+ "epoch": 36.055749128919864,
+ "grad_norm": 0.052498709410429,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 1.8275878429412842,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069686411149824,
+ "grad_norm": 0.05208680033683777,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 1.8087416887283325,
+ "step": 2597
+ },
+ {
+ "epoch": 36.08362369337979,
+ "grad_norm": 0.05346959829330444,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 1.8090479373931885,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09756097560975,
+ "grad_norm": 0.052402183413505554,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 1.8265495300292969,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11149825783972,
+ "grad_norm": 0.051748018711805344,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 1.8173667192459106,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12543554006969,
+ "grad_norm": 0.05207061767578125,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 1.821485996246338,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13937282229965,
+ "grad_norm": 0.05230815336108208,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 1.804908037185669,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153310104529616,
+ "grad_norm": 0.052593570202589035,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 1.8154864311218262,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167247386759584,
+ "grad_norm": 0.05259322747588158,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 1.8238556385040283,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181184668989545,
+ "grad_norm": 0.052652861922979355,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 1.8147733211517334,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19512195121951,
+ "grad_norm": 0.0530865415930748,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 1.8086535930633545,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20905923344948,
+ "grad_norm": 0.053423281759023666,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 1.8174736499786377,
+ "step": 2607
+ },
+ {
+ "epoch": 36.22299651567944,
+ "grad_norm": 0.0524580255150795,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 1.827101230621338,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23693379790941,
+ "grad_norm": 0.05376284569501877,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 1.8410698175430298,
+ "step": 2609
+ },
+ {
+ "epoch": 36.250871080139376,
+ "grad_norm": 0.05277293920516968,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 1.8262073993682861,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26480836236934,
+ "grad_norm": 0.05415027216076851,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 1.8311805725097656,
+ "step": 2611
+ },
+ {
+ "epoch": 36.278745644599304,
+ "grad_norm": 0.05142449960112572,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 1.816835880279541,
+ "step": 2612
+ },
+ {
+ "epoch": 36.292682926829265,
+ "grad_norm": 0.053121037781238556,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 1.818075180053711,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30662020905923,
+ "grad_norm": 0.051999714225530624,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 1.8257043361663818,
+ "step": 2614
+ },
+ {
+ "epoch": 36.3205574912892,
+ "grad_norm": 0.05257652327418327,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 1.814011573791504,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33449477351916,
+ "grad_norm": 0.0541016161441803,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 1.826779842376709,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34843205574913,
+ "grad_norm": 0.05318204313516617,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 1.8242204189300537,
+ "step": 2617
+ },
+ {
+ "epoch": 36.362369337979096,
+ "grad_norm": 0.05453426018357277,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 1.822621464729309,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37630662020906,
+ "grad_norm": 0.05376817658543587,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 1.8253248929977417,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390243902439025,
+ "grad_norm": 0.05483992025256157,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 1.8317890167236328,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40418118466899,
+ "grad_norm": 0.05439801886677742,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 1.835724949836731,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41811846689895,
+ "grad_norm": 0.055256184190511703,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 1.823319911956787,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43205574912892,
+ "grad_norm": 0.05287355184555054,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 1.8103458881378174,
+ "step": 2623
+ },
+ {
+ "epoch": 36.44599303135889,
+ "grad_norm": 0.05546991154551506,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 1.8309146165847778,
+ "step": 2624
+ },
+ {
+ "epoch": 36.45993031358885,
+ "grad_norm": 0.05322236195206642,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 1.821721076965332,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47386759581882,
+ "grad_norm": 0.054857514798641205,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 1.8251492977142334,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48780487804878,
+ "grad_norm": 0.053303562104701996,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 1.8257558345794678,
+ "step": 2627
+ },
+ {
+ "epoch": 36.501742160278745,
+ "grad_norm": 0.05594215914607048,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 1.835466742515564,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51567944250871,
+ "grad_norm": 0.05624304339289665,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 1.8186707496643066,
+ "step": 2629
+ },
+ {
+ "epoch": 36.52961672473867,
+ "grad_norm": 0.055144522339105606,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 1.8301169872283936,
+ "step": 2630
+ },
+ {
+ "epoch": 36.54355400696864,
+ "grad_norm": 0.0535074919462204,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 1.8138389587402344,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55749128919861,
+ "grad_norm": 0.05484950542449951,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 1.8223397731781006,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57142857142857,
+ "grad_norm": 0.05473049357533455,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 1.8218905925750732,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58536585365854,
+ "grad_norm": 0.05588974803686142,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 1.825666904449463,
+ "step": 2634
+ },
+ {
+ "epoch": 36.599303135888505,
+ "grad_norm": 0.05313410982489586,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 1.8338199853897095,
+ "step": 2635
+ },
+ {
+ "epoch": 36.613240418118465,
+ "grad_norm": 0.056002989411354065,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 1.8219728469848633,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62717770034843,
+ "grad_norm": 0.05519993603229523,
+ "learning_rate": 0.000110418175419276,
+ "loss": 1.826596736907959,
+ "step": 2637
+ },
+ {
+ "epoch": 36.641114982578394,
+ "grad_norm": 0.055067840963602066,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 1.8218865394592285,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65505226480836,
+ "grad_norm": 0.0548541434109211,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 1.8283402919769287,
+ "step": 2639
+ },
+ {
+ "epoch": 36.66898954703833,
+ "grad_norm": 0.05756586790084839,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 1.8259823322296143,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68292682926829,
+ "grad_norm": 0.05620390921831131,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 1.8149542808532715,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69686411149826,
+ "grad_norm": 0.05768077075481415,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 1.8275048732757568,
+ "step": 2642
+ },
+ {
+ "epoch": 36.710801393728225,
+ "grad_norm": 0.05531959980726242,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 1.8295093774795532,
+ "step": 2643
+ },
+ {
+ "epoch": 36.724738675958186,
+ "grad_norm": 0.057147517800331116,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 1.8217309713363647,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73867595818815,
+ "grad_norm": 0.05488502234220505,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 1.8418645858764648,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75261324041812,
+ "grad_norm": 0.05747298523783684,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 1.825081467628479,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76655052264808,
+ "grad_norm": 0.05621277540922165,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 1.8208179473876953,
+ "step": 2647
+ },
+ {
+ "epoch": 36.78048780487805,
+ "grad_norm": 0.05534190684556961,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 1.8269200325012207,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79442508710802,
+ "grad_norm": 0.0570702850818634,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 1.8310284614562988,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80836236933798,
+ "grad_norm": 0.055876970291137695,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 1.831892967224121,
+ "step": 2650
+ },
+ {
+ "epoch": 36.822299651567945,
+ "grad_norm": 0.05680249258875847,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 1.8280737400054932,
+ "step": 2651
+ },
+ {
+ "epoch": 36.836236933797906,
+ "grad_norm": 0.056308578699827194,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 1.844076156616211,
+ "step": 2652
+ },
+ {
+ "epoch": 36.850174216027874,
+ "grad_norm": 0.05752384290099144,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 1.8403749465942383,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86411149825784,
+ "grad_norm": 0.058181654661893845,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 1.8354277610778809,
+ "step": 2654
+ },
+ {
+ "epoch": 36.8780487804878,
+ "grad_norm": 0.05747664347290993,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 1.8315486907958984,
+ "step": 2655
+ },
+ {
+ "epoch": 36.89198606271777,
+ "grad_norm": 0.057498663663864136,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 1.8256535530090332,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90592334494774,
+ "grad_norm": 0.0565464086830616,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 1.8265036344528198,
+ "step": 2657
+ },
+ {
+ "epoch": 36.9198606271777,
+ "grad_norm": 0.057560283690690994,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 1.822776198387146,
+ "step": 2658
+ },
+ {
+ "epoch": 36.933797909407666,
+ "grad_norm": 0.05598663538694382,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 1.8307170867919922,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94773519163763,
+ "grad_norm": 0.0580892413854599,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 1.831838846206665,
+ "step": 2660
+ },
+ {
+ "epoch": 36.961672473867594,
+ "grad_norm": 0.056303758174180984,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 1.837568998336792,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97560975609756,
+ "grad_norm": 0.05781455338001251,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 1.8347523212432861,
+ "step": 2662
+ },
+ {
+ "epoch": 36.98954703832753,
+ "grad_norm": 0.05536883324384689,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 1.8196232318878174,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.046378862112760544,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 1.3591325283050537,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6459980607032776,
+ "eval_runtime": 40.631,
+ "eval_samples_per_second": 60.102,
+ "eval_steps_per_second": 0.492,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01393728222997,
+ "grad_norm": 0.050504088401794434,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 1.8197872638702393,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02787456445993,
+ "grad_norm": 0.05133911967277527,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 1.8092283010482788,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041811846689896,
+ "grad_norm": 0.051401812583208084,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 1.7972350120544434,
+ "step": 2667
+ },
+ {
+ "epoch": 37.055749128919864,
+ "grad_norm": 0.050701867789030075,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 1.8077106475830078,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069686411149824,
+ "grad_norm": 0.05219056457281113,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 1.7995078563690186,
+ "step": 2669
+ },
+ {
+ "epoch": 37.08362369337979,
+ "grad_norm": 0.05003709718585014,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 1.8166279792785645,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09756097560975,
+ "grad_norm": 0.051945704966783524,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 1.8085274696350098,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11149825783972,
+ "grad_norm": 0.05073661729693413,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 1.794724941253662,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12543554006969,
+ "grad_norm": 0.051837258040905,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 1.8101842403411865,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13937282229965,
+ "grad_norm": 0.051624711602926254,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 1.8128712177276611,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153310104529616,
+ "grad_norm": 0.05337493121623993,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 1.8105556964874268,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167247386759584,
+ "grad_norm": 0.05040125921368599,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 1.801112413406372,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181184668989545,
+ "grad_norm": 0.05222955718636513,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 1.8220586776733398,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19512195121951,
+ "grad_norm": 0.052356619387865067,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 1.8208140134811401,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20905923344948,
+ "grad_norm": 0.05221114307641983,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 1.8082315921783447,
+ "step": 2679
+ },
+ {
+ "epoch": 37.22299651567944,
+ "grad_norm": 0.05170206353068352,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 1.8159395456314087,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23693379790941,
+ "grad_norm": 0.05091318488121033,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 1.8138477802276611,
+ "step": 2681
+ },
+ {
+ "epoch": 37.250871080139376,
+ "grad_norm": 0.05228526517748833,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 1.8149440288543701,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26480836236934,
+ "grad_norm": 0.053375136107206345,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 1.8169589042663574,
+ "step": 2683
+ },
+ {
+ "epoch": 37.278745644599304,
+ "grad_norm": 0.05215807631611824,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 1.8025376796722412,
+ "step": 2684
+ },
+ {
+ "epoch": 37.292682926829265,
+ "grad_norm": 0.05166545882821083,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 1.815457820892334,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30662020905923,
+ "grad_norm": 0.05314334109425545,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 1.8092843294143677,
+ "step": 2686
+ },
+ {
+ "epoch": 37.3205574912892,
+ "grad_norm": 0.0525335855782032,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 1.8068681955337524,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33449477351916,
+ "grad_norm": 0.05154058337211609,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 1.8220410346984863,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34843205574913,
+ "grad_norm": 0.05400070548057556,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 1.8106049299240112,
+ "step": 2689
+ },
+ {
+ "epoch": 37.362369337979096,
+ "grad_norm": 0.051827941089868546,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 1.8139021396636963,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37630662020906,
+ "grad_norm": 0.053685542196035385,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 1.8051722049713135,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390243902439025,
+ "grad_norm": 0.05291858687996864,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 1.806849479675293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40418118466899,
+ "grad_norm": 0.05310729518532753,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 1.7940170764923096,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41811846689895,
+ "grad_norm": 0.05288896709680557,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 1.824284553527832,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43205574912892,
+ "grad_norm": 0.052886299788951874,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 1.8206238746643066,
+ "step": 2695
+ },
+ {
+ "epoch": 37.44599303135889,
+ "grad_norm": 0.05356050282716751,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 1.8159115314483643,
+ "step": 2696
+ },
+ {
+ "epoch": 37.45993031358885,
+ "grad_norm": 0.053028833121061325,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 1.8148610591888428,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47386759581882,
+ "grad_norm": 0.05355606600642204,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 1.8167890310287476,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48780487804878,
+ "grad_norm": 0.053764909505844116,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 1.8095712661743164,
+ "step": 2699
+ },
+ {
+ "epoch": 37.501742160278745,
+ "grad_norm": 0.05224880203604698,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 1.8151801824569702,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51567944250871,
+ "grad_norm": 0.05442487820982933,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 1.8090888261795044,
+ "step": 2701
+ },
+ {
+ "epoch": 37.52961672473867,
+ "grad_norm": 0.05287255719304085,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 1.8190618753433228,
+ "step": 2702
+ },
+ {
+ "epoch": 37.54355400696864,
+ "grad_norm": 0.05432182550430298,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 1.8103792667388916,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55749128919861,
+ "grad_norm": 0.05345293879508972,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 1.8107848167419434,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57142857142857,
+ "grad_norm": 0.05567241832613945,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 1.822261095046997,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58536585365854,
+ "grad_norm": 0.0534495934844017,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 1.8195551633834839,
+ "step": 2706
+ },
+ {
+ "epoch": 37.599303135888505,
+ "grad_norm": 0.05432790517807007,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 1.8119161128997803,
+ "step": 2707
+ },
+ {
+ "epoch": 37.613240418118465,
+ "grad_norm": 0.05465250089764595,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 1.8269178867340088,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62717770034843,
+ "grad_norm": 0.05540885776281357,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 1.8170043230056763,
+ "step": 2709
+ },
+ {
+ "epoch": 37.641114982578394,
+ "grad_norm": 0.053956300020217896,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 1.8141452074050903,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65505226480836,
+ "grad_norm": 0.0530884824693203,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 1.8228175640106201,
+ "step": 2711
+ },
+ {
+ "epoch": 37.66898954703833,
+ "grad_norm": 0.05414469167590141,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 1.8148912191390991,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68292682926829,
+ "grad_norm": 0.054135460406541824,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 1.8189573287963867,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69686411149826,
+ "grad_norm": 0.05427670478820801,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 1.8352646827697754,
+ "step": 2714
+ },
+ {
+ "epoch": 37.710801393728225,
+ "grad_norm": 0.054425179958343506,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 1.8101688623428345,
+ "step": 2715
+ },
+ {
+ "epoch": 37.724738675958186,
+ "grad_norm": 0.05361952632665634,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 1.822049856185913,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73867595818815,
+ "grad_norm": 0.053644467145204544,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 1.8222166299819946,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75261324041812,
+ "grad_norm": 0.054326146841049194,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 1.8122522830963135,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76655052264808,
+ "grad_norm": 0.05346079170703888,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 1.8157234191894531,
+ "step": 2719
+ },
+ {
+ "epoch": 37.78048780487805,
+ "grad_norm": 0.05502757802605629,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 1.8146154880523682,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79442508710802,
+ "grad_norm": 0.05353285372257233,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 1.81754732131958,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80836236933798,
+ "grad_norm": 0.05462120845913887,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 1.8201050758361816,
+ "step": 2722
+ },
+ {
+ "epoch": 37.822299651567945,
+ "grad_norm": 0.05272475257515907,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 1.8173773288726807,
+ "step": 2723
+ },
+ {
+ "epoch": 37.836236933797906,
+ "grad_norm": 0.056003544479608536,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 1.8211115598678589,
+ "step": 2724
+ },
+ {
+ "epoch": 37.850174216027874,
+ "grad_norm": 0.05486152321100235,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 1.8221628665924072,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86411149825784,
+ "grad_norm": 0.05549563467502594,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 1.8200361728668213,
+ "step": 2726
+ },
+ {
+ "epoch": 37.8780487804878,
+ "grad_norm": 0.05407272279262543,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 1.814354658126831,
+ "step": 2727
+ },
+ {
+ "epoch": 37.89198606271777,
+ "grad_norm": 0.05526091903448105,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 1.8025412559509277,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90592334494774,
+ "grad_norm": 0.05579185113310814,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 1.8137646913528442,
+ "step": 2729
+ },
+ {
+ "epoch": 37.9198606271777,
+ "grad_norm": 0.054667308926582336,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 1.8100636005401611,
+ "step": 2730
+ },
+ {
+ "epoch": 37.933797909407666,
+ "grad_norm": 0.054524678736925125,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 1.8058973550796509,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94773519163763,
+ "grad_norm": 0.05454818904399872,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 1.8153016567230225,
+ "step": 2732
+ },
+ {
+ "epoch": 37.961672473867594,
+ "grad_norm": 0.05491559952497482,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 1.8142199516296387,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97560975609756,
+ "grad_norm": 0.055562835186719894,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 1.8082599639892578,
+ "step": 2734
+ },
+ {
+ "epoch": 37.98954703832753,
+ "grad_norm": 0.055928416550159454,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 1.8185434341430664,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.044970303773880005,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 1.346644401550293,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6474234461784363,
+ "eval_runtime": 40.6464,
+ "eval_samples_per_second": 60.079,
+ "eval_steps_per_second": 0.492,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01393728222997,
+ "grad_norm": 0.05179993808269501,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 1.7899185419082642,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02787456445993,
+ "grad_norm": 0.04930533468723297,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 1.7904696464538574,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041811846689896,
+ "grad_norm": 0.052794501185417175,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 1.7964375019073486,
+ "step": 2739
+ },
+ {
+ "epoch": 38.055749128919864,
+ "grad_norm": 0.049202270805835724,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 1.790614128112793,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069686411149824,
+ "grad_norm": 0.04903653636574745,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 1.803176999092102,
+ "step": 2741
+ },
+ {
+ "epoch": 38.08362369337979,
+ "grad_norm": 0.04976186528801918,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 1.8021293878555298,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09756097560975,
+ "grad_norm": 0.05226270109415054,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 1.7997218370437622,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11149825783972,
+ "grad_norm": 0.04967877268791199,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 1.8076910972595215,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12543554006969,
+ "grad_norm": 0.05012716352939606,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 1.7991023063659668,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13937282229965,
+ "grad_norm": 0.05108457803726196,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 1.8080018758773804,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153310104529616,
+ "grad_norm": 0.05005490034818649,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 1.7945563793182373,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167247386759584,
+ "grad_norm": 0.050113365054130554,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 1.802520990371704,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181184668989545,
+ "grad_norm": 0.05096133425831795,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 1.7994227409362793,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19512195121951,
+ "grad_norm": 0.05119607597589493,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 1.8115901947021484,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20905923344948,
+ "grad_norm": 0.051627811044454575,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 1.800724744796753,
+ "step": 2751
+ },
+ {
+ "epoch": 38.22299651567944,
+ "grad_norm": 0.05053049698472023,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 1.7860815525054932,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23693379790941,
+ "grad_norm": 0.05122068524360657,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 1.7976000308990479,
+ "step": 2753
+ },
+ {
+ "epoch": 38.250871080139376,
+ "grad_norm": 0.05197495222091675,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 1.811307430267334,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26480836236934,
+ "grad_norm": 0.051113326102495193,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 1.7974506616592407,
+ "step": 2755
+ },
+ {
+ "epoch": 38.278745644599304,
+ "grad_norm": 0.05044601857662201,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 1.8004801273345947,
+ "step": 2756
+ },
+ {
+ "epoch": 38.292682926829265,
+ "grad_norm": 0.05080671235918999,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 1.7960381507873535,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30662020905923,
+ "grad_norm": 0.05112270265817642,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 1.8070579767227173,
+ "step": 2758
+ },
+ {
+ "epoch": 38.3205574912892,
+ "grad_norm": 0.050059717148542404,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 1.7989699840545654,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33449477351916,
+ "grad_norm": 0.05151006579399109,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 1.793050765991211,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34843205574913,
+ "grad_norm": 0.05045556277036667,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 1.7945470809936523,
+ "step": 2761
+ },
+ {
+ "epoch": 38.362369337979096,
+ "grad_norm": 0.05152078717947006,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 1.8155558109283447,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37630662020906,
+ "grad_norm": 0.049696408212184906,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 1.7929422855377197,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390243902439025,
+ "grad_norm": 0.050947509706020355,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 1.8029417991638184,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40418118466899,
+ "grad_norm": 0.04988504573702812,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 1.807220697402954,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41811846689895,
+ "grad_norm": 0.05031213164329529,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 1.820298433303833,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43205574912892,
+ "grad_norm": 0.05067938566207886,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 1.8025063276290894,
+ "step": 2767
+ },
+ {
+ "epoch": 38.44599303135889,
+ "grad_norm": 0.050599709153175354,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 1.7912819385528564,
+ "step": 2768
+ },
+ {
+ "epoch": 38.45993031358885,
+ "grad_norm": 0.05000261962413788,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 1.8116015195846558,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47386759581882,
+ "grad_norm": 0.05164877697825432,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 1.7894442081451416,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48780487804878,
+ "grad_norm": 0.05037420988082886,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 1.8002345561981201,
+ "step": 2771
+ },
+ {
+ "epoch": 38.501742160278745,
+ "grad_norm": 0.05061398074030876,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 1.8103997707366943,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51567944250871,
+ "grad_norm": 0.05141839757561684,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 1.7920000553131104,
+ "step": 2773
+ },
+ {
+ "epoch": 38.52961672473867,
+ "grad_norm": 0.050547219812870026,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 1.8011581897735596,
+ "step": 2774
+ },
+ {
+ "epoch": 38.54355400696864,
+ "grad_norm": 0.052665967494249344,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 1.8171170949935913,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55749128919861,
+ "grad_norm": 0.05148577317595482,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 1.7976020574569702,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57142857142857,
+ "grad_norm": 0.05378367006778717,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 1.8061752319335938,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58536585365854,
+ "grad_norm": 0.05071459710597992,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 1.8084588050842285,
+ "step": 2778
+ },
+ {
+ "epoch": 38.599303135888505,
+ "grad_norm": 0.05229831859469414,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 1.8100000619888306,
+ "step": 2779
+ },
+ {
+ "epoch": 38.613240418118465,
+ "grad_norm": 0.052434664219617844,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 1.8119068145751953,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62717770034843,
+ "grad_norm": 0.05269220098853111,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 1.799889087677002,
+ "step": 2781
+ },
+ {
+ "epoch": 38.641114982578394,
+ "grad_norm": 0.05271878093481064,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 1.792294979095459,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65505226480836,
+ "grad_norm": 0.05291058123111725,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 1.7999838590621948,
+ "step": 2783
+ },
+ {
+ "epoch": 38.66898954703833,
+ "grad_norm": 0.05346683785319328,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 1.801847219467163,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68292682926829,
+ "grad_norm": 0.053118057548999786,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 1.7951825857162476,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69686411149826,
+ "grad_norm": 0.0527576319873333,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 1.8107949495315552,
+ "step": 2786
+ },
+ {
+ "epoch": 38.710801393728225,
+ "grad_norm": 0.05188573896884918,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 1.8066956996917725,
+ "step": 2787
+ },
+ {
+ "epoch": 38.724738675958186,
+ "grad_norm": 0.05341636762022972,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 1.8036713600158691,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73867595818815,
+ "grad_norm": 0.051814883947372437,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 1.8106606006622314,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75261324041812,
+ "grad_norm": 0.05374172329902649,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 1.7904407978057861,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76655052264808,
+ "grad_norm": 0.052160196006298065,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 1.797128677368164,
+ "step": 2791
+ },
+ {
+ "epoch": 38.78048780487805,
+ "grad_norm": 0.052794717252254486,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 1.8085696697235107,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79442508710802,
+ "grad_norm": 0.05286112427711487,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 1.8027613162994385,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80836236933798,
+ "grad_norm": 0.052660733461380005,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 1.8135063648223877,
+ "step": 2794
+ },
+ {
+ "epoch": 38.822299651567945,
+ "grad_norm": 0.052824389189481735,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 1.7999730110168457,
+ "step": 2795
+ },
+ {
+ "epoch": 38.836236933797906,
+ "grad_norm": 0.05390750244259834,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 1.7879369258880615,
+ "step": 2796
+ },
+ {
+ "epoch": 38.850174216027874,
+ "grad_norm": 0.053744010627269745,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 1.811363935470581,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86411149825784,
+ "grad_norm": 0.05401702970266342,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 1.8049407005310059,
+ "step": 2798
+ },
+ {
+ "epoch": 38.8780487804878,
+ "grad_norm": 0.05376699939370155,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 1.8111652135849,
+ "step": 2799
+ },
+ {
+ "epoch": 38.89198606271777,
+ "grad_norm": 0.05381734296679497,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 1.8312594890594482,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90592334494774,
+ "grad_norm": 0.05387615039944649,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 1.805098295211792,
+ "step": 2801
+ },
+ {
+ "epoch": 38.9198606271777,
+ "grad_norm": 0.05285770446062088,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 1.8160698413848877,
+ "step": 2802
+ },
+ {
+ "epoch": 38.933797909407666,
+ "grad_norm": 0.05545372888445854,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 1.8070565462112427,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94773519163763,
+ "grad_norm": 0.05404449254274368,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 1.8055229187011719,
+ "step": 2804
+ },
+ {
+ "epoch": 38.961672473867594,
+ "grad_norm": 0.05500908941030502,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 1.798984169960022,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97560975609756,
+ "grad_norm": 0.05408451333642006,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 1.8082829713821411,
+ "step": 2806
+ },
+ {
+ "epoch": 38.98954703832753,
+ "grad_norm": 0.0534386970102787,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 1.8156507015228271,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.044039830565452576,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 1.3510627746582031,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6490727663040161,
+ "eval_runtime": 41.0373,
+ "eval_samples_per_second": 59.507,
+ "eval_steps_per_second": 0.487,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01393728222997,
+ "grad_norm": 0.0485052689909935,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 1.787034511566162,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02787456445993,
+ "grad_norm": 0.046773314476013184,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 1.7824962139129639,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041811846689896,
+ "grad_norm": 0.04776127636432648,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 1.784999132156372,
+ "step": 2811
+ },
+ {
+ "epoch": 39.055749128919864,
+ "grad_norm": 0.04702889546751976,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 1.7887163162231445,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069686411149824,
+ "grad_norm": 0.0472981221973896,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 1.7778058052062988,
+ "step": 2813
+ },
+ {
+ "epoch": 39.08362369337979,
+ "grad_norm": 0.04829077050089836,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 1.8164949417114258,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09756097560975,
+ "grad_norm": 0.04883906990289688,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 1.787117838859558,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11149825783972,
+ "grad_norm": 0.0464581660926342,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 1.7883102893829346,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12543554006969,
+ "grad_norm": 0.047438278794288635,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 1.7923827171325684,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13937282229965,
+ "grad_norm": 0.04703788831830025,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 1.78762686252594,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153310104529616,
+ "grad_norm": 0.049375683069229126,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 1.7970117330551147,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167247386759584,
+ "grad_norm": 0.04796822741627693,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 1.778883457183838,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181184668989545,
+ "grad_norm": 0.04719251021742821,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 1.7749544382095337,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19512195121951,
+ "grad_norm": 0.04893529415130615,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 1.787841796875,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20905923344948,
+ "grad_norm": 0.04869271442294121,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 1.8007760047912598,
+ "step": 2823
+ },
+ {
+ "epoch": 39.22299651567944,
+ "grad_norm": 0.04950389638543129,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 1.7905025482177734,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23693379790941,
+ "grad_norm": 0.0489855520427227,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 1.7977821826934814,
+ "step": 2825
+ },
+ {
+ "epoch": 39.250871080139376,
+ "grad_norm": 0.04881766065955162,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 1.7955210208892822,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26480836236934,
+ "grad_norm": 0.04834857955574989,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 1.7897529602050781,
+ "step": 2827
+ },
+ {
+ "epoch": 39.278745644599304,
+ "grad_norm": 0.04898613691329956,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 1.7888939380645752,
+ "step": 2828
+ },
+ {
+ "epoch": 39.292682926829265,
+ "grad_norm": 0.04936131089925766,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 1.8096022605895996,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30662020905923,
+ "grad_norm": 0.04834993928670883,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 1.800262689590454,
+ "step": 2830
+ },
+ {
+ "epoch": 39.3205574912892,
+ "grad_norm": 0.049713220447301865,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 1.7950725555419922,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33449477351916,
+ "grad_norm": 0.049898866564035416,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 1.7812716960906982,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34843205574913,
+ "grad_norm": 0.049175627529621124,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 1.7746937274932861,
+ "step": 2833
+ },
+ {
+ "epoch": 39.362369337979096,
+ "grad_norm": 0.04872290790081024,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 1.7824972867965698,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37630662020906,
+ "grad_norm": 0.049543045461177826,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 1.793465256690979,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390243902439025,
+ "grad_norm": 0.049746304750442505,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 1.792328119277954,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40418118466899,
+ "grad_norm": 0.04971722885966301,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 1.7932684421539307,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41811846689895,
+ "grad_norm": 0.05128713324666023,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 1.8031436204910278,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43205574912892,
+ "grad_norm": 0.04928988218307495,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 1.7928621768951416,
+ "step": 2839
+ },
+ {
+ "epoch": 39.44599303135889,
+ "grad_norm": 0.05211673676967621,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 1.7749731540679932,
+ "step": 2840
+ },
+ {
+ "epoch": 39.45993031358885,
+ "grad_norm": 0.05039450526237488,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 1.7910631895065308,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47386759581882,
+ "grad_norm": 0.05047272890806198,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 1.8027098178863525,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48780487804878,
+ "grad_norm": 0.05042344331741333,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 1.8134506940841675,
+ "step": 2843
+ },
+ {
+ "epoch": 39.501742160278745,
+ "grad_norm": 0.0506373755633831,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 1.7785459756851196,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51567944250871,
+ "grad_norm": 0.04902270808815956,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 1.7985787391662598,
+ "step": 2845
+ },
+ {
+ "epoch": 39.52961672473867,
+ "grad_norm": 0.05162668600678444,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 1.7813456058502197,
+ "step": 2846
+ },
+ {
+ "epoch": 39.54355400696864,
+ "grad_norm": 0.05005355924367905,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 1.791803240776062,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55749128919861,
+ "grad_norm": 0.050947558134794235,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 1.7949843406677246,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57142857142857,
+ "grad_norm": 0.05130480229854584,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 1.7962243556976318,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58536585365854,
+ "grad_norm": 0.049898259341716766,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 1.7940798997879028,
+ "step": 2850
+ },
+ {
+ "epoch": 39.599303135888505,
+ "grad_norm": 0.050370268523693085,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 1.8001989126205444,
+ "step": 2851
+ },
+ {
+ "epoch": 39.613240418118465,
+ "grad_norm": 0.05121270939707756,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 1.79237961769104,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62717770034843,
+ "grad_norm": 0.05014092102646828,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 1.804434895515442,
+ "step": 2853
+ },
+ {
+ "epoch": 39.641114982578394,
+ "grad_norm": 0.050814609974622726,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 1.7997581958770752,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65505226480836,
+ "grad_norm": 0.05023293197154999,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 1.8026623725891113,
+ "step": 2855
+ },
+ {
+ "epoch": 39.66898954703833,
+ "grad_norm": 0.051415517926216125,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 1.8061354160308838,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68292682926829,
+ "grad_norm": 0.04993635416030884,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 1.7835321426391602,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69686411149826,
+ "grad_norm": 0.05157633125782013,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 1.8007519245147705,
+ "step": 2858
+ },
+ {
+ "epoch": 39.710801393728225,
+ "grad_norm": 0.050889477133750916,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 1.7806460857391357,
+ "step": 2859
+ },
+ {
+ "epoch": 39.724738675958186,
+ "grad_norm": 0.050646744668483734,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 1.780684471130371,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73867595818815,
+ "grad_norm": 0.05104859918355942,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 1.796810507774353,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75261324041812,
+ "grad_norm": 0.051958777010440826,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 1.8104615211486816,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76655052264808,
+ "grad_norm": 0.05137504264712334,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 1.7988557815551758,
+ "step": 2863
+ },
+ {
+ "epoch": 39.78048780487805,
+ "grad_norm": 0.051518820226192474,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 1.7865259647369385,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79442508710802,
+ "grad_norm": 0.051753368228673935,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 1.8039045333862305,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80836236933798,
+ "grad_norm": 0.0514046885073185,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 1.8040326833724976,
+ "step": 2866
+ },
+ {
+ "epoch": 39.822299651567945,
+ "grad_norm": 0.05259102210402489,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 1.8034707307815552,
+ "step": 2867
+ },
+ {
+ "epoch": 39.836236933797906,
+ "grad_norm": 0.05125397443771362,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 1.7860667705535889,
+ "step": 2868
+ },
+ {
+ "epoch": 39.850174216027874,
+ "grad_norm": 0.05120245739817619,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 1.7920024394989014,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86411149825784,
+ "grad_norm": 0.05269347131252289,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 1.7958977222442627,
+ "step": 2870
+ },
+ {
+ "epoch": 39.8780487804878,
+ "grad_norm": 0.050947729498147964,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 1.7864038944244385,
+ "step": 2871
+ },
+ {
+ "epoch": 39.89198606271777,
+ "grad_norm": 0.052531689405441284,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 1.795427918434143,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90592334494774,
+ "grad_norm": 0.052431896328926086,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 1.7928128242492676,
+ "step": 2873
+ },
+ {
+ "epoch": 39.9198606271777,
+ "grad_norm": 0.053628548979759216,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 1.7986962795257568,
+ "step": 2874
+ },
+ {
+ "epoch": 39.933797909407666,
+ "grad_norm": 0.05186983942985535,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 1.7967369556427002,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94773519163763,
+ "grad_norm": 0.05122850835323334,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 1.7836300134658813,
+ "step": 2876
+ },
+ {
+ "epoch": 39.961672473867594,
+ "grad_norm": 0.05265440791845322,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 1.801450490951538,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97560975609756,
+ "grad_norm": 0.053320810198783875,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 1.792264461517334,
+ "step": 2878
+ },
+ {
+ "epoch": 39.98954703832753,
+ "grad_norm": 0.05239804461598396,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 1.8015801906585693,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04351874440908432,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 1.3430795669555664,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6505019068717957,
+ "eval_runtime": 40.2814,
+ "eval_samples_per_second": 60.623,
+ "eval_steps_per_second": 0.497,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01393728222997,
+ "grad_norm": 0.04842356964945793,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 1.7894266843795776,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02787456445993,
+ "grad_norm": 0.04759506508708,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 1.7896207571029663,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041811846689896,
+ "grad_norm": 0.04793368652462959,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 1.7770864963531494,
+ "step": 2883
+ },
+ {
+ "epoch": 40.055749128919864,
+ "grad_norm": 0.04775947704911232,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 1.7689225673675537,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069686411149824,
+ "grad_norm": 0.04841279238462448,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 1.790536880493164,
+ "step": 2885
+ },
+ {
+ "epoch": 40.08362369337979,
+ "grad_norm": 0.04589880257844925,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 1.7804309129714966,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09756097560975,
+ "grad_norm": 0.04910781979560852,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 1.7737990617752075,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11149825783972,
+ "grad_norm": 0.04736403003334999,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 1.7843401432037354,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12543554006969,
+ "grad_norm": 0.047884710133075714,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 1.7776315212249756,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13937282229965,
+ "grad_norm": 0.047823768109083176,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 1.7916157245635986,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153310104529616,
+ "grad_norm": 0.04811791703104973,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 1.8077316284179688,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167247386759584,
+ "grad_norm": 0.05039852485060692,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 1.7821904420852661,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181184668989545,
+ "grad_norm": 0.04656902328133583,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 1.775067687034607,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19512195121951,
+ "grad_norm": 0.04947083443403244,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 1.7826871871948242,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20905923344948,
+ "grad_norm": 0.04797012731432915,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 1.792678713798523,
+ "step": 2895
+ },
+ {
+ "epoch": 40.22299651567944,
+ "grad_norm": 0.048327650874853134,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 1.7605221271514893,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23693379790941,
+ "grad_norm": 0.049527380615472794,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 1.7833847999572754,
+ "step": 2897
+ },
+ {
+ "epoch": 40.250871080139376,
+ "grad_norm": 0.04733026400208473,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 1.7732503414154053,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26480836236934,
+ "grad_norm": 0.04898470640182495,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 1.776735544204712,
+ "step": 2899
+ },
+ {
+ "epoch": 40.278745644599304,
+ "grad_norm": 0.04847757890820503,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 1.7859764099121094,
+ "step": 2900
+ },
+ {
+ "epoch": 40.292682926829265,
+ "grad_norm": 0.04857335984706879,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 1.7744362354278564,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30662020905923,
+ "grad_norm": 0.047604888677597046,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 1.7872700691223145,
+ "step": 2902
+ },
+ {
+ "epoch": 40.3205574912892,
+ "grad_norm": 0.048553720116615295,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 1.7771832942962646,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33449477351916,
+ "grad_norm": 0.047940365970134735,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 1.7938380241394043,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34843205574913,
+ "grad_norm": 0.04888467490673065,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 1.7793422937393188,
+ "step": 2905
+ },
+ {
+ "epoch": 40.362369337979096,
+ "grad_norm": 0.04792848229408264,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 1.7843034267425537,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37630662020906,
+ "grad_norm": 0.04912486672401428,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 1.7690515518188477,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390243902439025,
+ "grad_norm": 0.04807637259364128,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 1.7809380292892456,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40418118466899,
+ "grad_norm": 0.04979358986020088,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 1.783470869064331,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41811846689895,
+ "grad_norm": 0.049359280616045,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 1.7851728200912476,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43205574912892,
+ "grad_norm": 0.04928767681121826,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 1.7869231700897217,
+ "step": 2911
+ },
+ {
+ "epoch": 40.44599303135889,
+ "grad_norm": 0.04966576769948006,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 1.771344542503357,
+ "step": 2912
+ },
+ {
+ "epoch": 40.45993031358885,
+ "grad_norm": 0.04966041445732117,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 1.7944121360778809,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47386759581882,
+ "grad_norm": 0.04922773316502571,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 1.798032522201538,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48780487804878,
+ "grad_norm": 0.04820060357451439,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 1.7730165719985962,
+ "step": 2915
+ },
+ {
+ "epoch": 40.501742160278745,
+ "grad_norm": 0.04986457899212837,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 1.7808765172958374,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51567944250871,
+ "grad_norm": 0.04911157861351967,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 1.7903361320495605,
+ "step": 2917
+ },
+ {
+ "epoch": 40.52961672473867,
+ "grad_norm": 0.048055119812488556,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 1.7886831760406494,
+ "step": 2918
+ },
+ {
+ "epoch": 40.54355400696864,
+ "grad_norm": 0.049517903476953506,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 1.7857229709625244,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55749128919861,
+ "grad_norm": 0.048604514449834824,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 1.7922042608261108,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57142857142857,
+ "grad_norm": 0.04886873811483383,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 1.7781345844268799,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58536585365854,
+ "grad_norm": 0.04913982376456261,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 1.7909157276153564,
+ "step": 2922
+ },
+ {
+ "epoch": 40.599303135888505,
+ "grad_norm": 0.04856523871421814,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 1.7691776752471924,
+ "step": 2923
+ },
+ {
+ "epoch": 40.613240418118465,
+ "grad_norm": 0.05121750012040138,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 1.7849647998809814,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62717770034843,
+ "grad_norm": 0.04991038516163826,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 1.7836909294128418,
+ "step": 2925
+ },
+ {
+ "epoch": 40.641114982578394,
+ "grad_norm": 0.04965447261929512,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 1.7835614681243896,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65505226480836,
+ "grad_norm": 0.049823079258203506,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 1.7881591320037842,
+ "step": 2927
+ },
+ {
+ "epoch": 40.66898954703833,
+ "grad_norm": 0.049288421869277954,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 1.7872439622879028,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68292682926829,
+ "grad_norm": 0.05006309971213341,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 1.776764988899231,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69686411149826,
+ "grad_norm": 0.05022395774722099,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 1.8023052215576172,
+ "step": 2930
+ },
+ {
+ "epoch": 40.710801393728225,
+ "grad_norm": 0.050583839416503906,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 1.783090353012085,
+ "step": 2931
+ },
+ {
+ "epoch": 40.724738675958186,
+ "grad_norm": 0.04937582463026047,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 1.7982478141784668,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73867595818815,
+ "grad_norm": 0.04985835775732994,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 1.7760167121887207,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75261324041812,
+ "grad_norm": 0.05028945952653885,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 1.7793045043945312,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76655052264808,
+ "grad_norm": 0.05060959234833717,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 1.7912629842758179,
+ "step": 2935
+ },
+ {
+ "epoch": 40.78048780487805,
+ "grad_norm": 0.05209024250507355,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 1.7781007289886475,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79442508710802,
+ "grad_norm": 0.04997926205396652,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 1.7866160869598389,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80836236933798,
+ "grad_norm": 0.05141197517514229,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 1.7921373844146729,
+ "step": 2938
+ },
+ {
+ "epoch": 40.822299651567945,
+ "grad_norm": 0.05031934380531311,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 1.786409854888916,
+ "step": 2939
+ },
+ {
+ "epoch": 40.836236933797906,
+ "grad_norm": 0.049375422298908234,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 1.7676091194152832,
+ "step": 2940
+ },
+ {
+ "epoch": 40.850174216027874,
+ "grad_norm": 0.05016583576798439,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 1.7917009592056274,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86411149825784,
+ "grad_norm": 0.04995865747332573,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 1.8048253059387207,
+ "step": 2942
+ },
+ {
+ "epoch": 40.8780487804878,
+ "grad_norm": 0.050674330443143845,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 1.7829941511154175,
+ "step": 2943
+ },
+ {
+ "epoch": 40.89198606271777,
+ "grad_norm": 0.05021820589900017,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 1.7910025119781494,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90592334494774,
+ "grad_norm": 0.05094406381249428,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 1.7892206907272339,
+ "step": 2945
+ },
+ {
+ "epoch": 40.9198606271777,
+ "grad_norm": 0.05145988613367081,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 1.783717155456543,
+ "step": 2946
+ },
+ {
+ "epoch": 40.933797909407666,
+ "grad_norm": 0.05058317258954048,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 1.7758351564407349,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94773519163763,
+ "grad_norm": 0.0512106828391552,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 1.8036705255508423,
+ "step": 2948
+ },
+ {
+ "epoch": 40.961672473867594,
+ "grad_norm": 0.0501241534948349,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 1.7745720148086548,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97560975609756,
+ "grad_norm": 0.05166011303663254,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 1.7932977676391602,
+ "step": 2950
+ },
+ {
+ "epoch": 40.98954703832753,
+ "grad_norm": 0.05050530284643173,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 1.7823102474212646,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.04222384840250015,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 1.3391203880310059,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.6519365310668945,
+ "eval_runtime": 40.5579,
+ "eval_samples_per_second": 60.21,
+ "eval_steps_per_second": 0.493,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01393728222997,
+ "grad_norm": 0.04689764603972435,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 1.7729463577270508,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02787456445993,
+ "grad_norm": 0.044540271162986755,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 1.7623910903930664,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041811846689896,
+ "grad_norm": 0.04553588107228279,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 1.7855658531188965,
+ "step": 2955
+ },
+ {
+ "epoch": 41.055749128919864,
+ "grad_norm": 0.04392184317111969,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 1.7711176872253418,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069686411149824,
+ "grad_norm": 0.04417131096124649,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 1.7741881608963013,
+ "step": 2957
+ },
+ {
+ "epoch": 41.08362369337979,
+ "grad_norm": 0.04466690495610237,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 1.7754600048065186,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09756097560975,
+ "grad_norm": 0.045785821974277496,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 1.7923266887664795,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11149825783972,
+ "grad_norm": 0.044646237045526505,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 1.768218994140625,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12543554006969,
+ "grad_norm": 0.04554557055234909,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 1.7587878704071045,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13937282229965,
+ "grad_norm": 0.0454886369407177,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 1.7782719135284424,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153310104529616,
+ "grad_norm": 0.045426227152347565,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 1.775224208831787,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167247386759584,
+ "grad_norm": 0.0466192252933979,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 1.7643415927886963,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181184668989545,
+ "grad_norm": 0.045661959797143936,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 1.7602108716964722,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19512195121951,
+ "grad_norm": 0.04575974866747856,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 1.7654622793197632,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20905923344948,
+ "grad_norm": 0.04690549150109291,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 1.7760295867919922,
+ "step": 2967
+ },
+ {
+ "epoch": 41.22299651567944,
+ "grad_norm": 0.04673248901963234,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 1.7847099304199219,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23693379790941,
+ "grad_norm": 0.04652085155248642,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 1.7894890308380127,
+ "step": 2969
+ },
+ {
+ "epoch": 41.250871080139376,
+ "grad_norm": 0.04779936745762825,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 1.7647104263305664,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26480836236934,
+ "grad_norm": 0.04592272266745567,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 1.7805603742599487,
+ "step": 2971
+ },
+ {
+ "epoch": 41.278745644599304,
+ "grad_norm": 0.0465652234852314,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 1.7796931266784668,
+ "step": 2972
+ },
+ {
+ "epoch": 41.292682926829265,
+ "grad_norm": 0.04658602178096771,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 1.7612698078155518,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30662020905923,
+ "grad_norm": 0.04615809768438339,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 1.7831511497497559,
+ "step": 2974
+ },
+ {
+ "epoch": 41.3205574912892,
+ "grad_norm": 0.04704147204756737,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 1.7652500867843628,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33449477351916,
+ "grad_norm": 0.046908169984817505,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 1.7832311391830444,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34843205574913,
+ "grad_norm": 0.046408504247665405,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 1.7829337120056152,
+ "step": 2977
+ },
+ {
+ "epoch": 41.362369337979096,
+ "grad_norm": 0.04615022614598274,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 1.758888840675354,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37630662020906,
+ "grad_norm": 0.046176742762327194,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 1.7705339193344116,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390243902439025,
+ "grad_norm": 0.048181675374507904,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 1.7633423805236816,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40418118466899,
+ "grad_norm": 0.047566913068294525,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 1.7715733051300049,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41811846689895,
+ "grad_norm": 0.046077970415353775,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 1.7751054763793945,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43205574912892,
+ "grad_norm": 0.045808132737874985,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 1.7841391563415527,
+ "step": 2983
+ },
+ {
+ "epoch": 41.44599303135889,
+ "grad_norm": 0.04701954126358032,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 1.7742592096328735,
+ "step": 2984
+ },
+ {
+ "epoch": 41.45993031358885,
+ "grad_norm": 0.04649316146969795,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 1.7765004634857178,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47386759581882,
+ "grad_norm": 0.046931155025959015,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 1.780288815498352,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48780487804878,
+ "grad_norm": 0.04566241428256035,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 1.781214952468872,
+ "step": 2987
+ },
+ {
+ "epoch": 41.501742160278745,
+ "grad_norm": 0.04714824631810188,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 1.77776038646698,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51567944250871,
+ "grad_norm": 0.046437617391347885,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 1.764427900314331,
+ "step": 2989
+ },
+ {
+ "epoch": 41.52961672473867,
+ "grad_norm": 0.047795433551073074,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 1.7826989889144897,
+ "step": 2990
+ },
+ {
+ "epoch": 41.54355400696864,
+ "grad_norm": 0.047585342079401016,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 1.791701316833496,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55749128919861,
+ "grad_norm": 0.046824004501104355,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 1.781235933303833,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57142857142857,
+ "grad_norm": 0.04806610569357872,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 1.7798388004302979,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58536585365854,
+ "grad_norm": 0.04673585668206215,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 1.7777440547943115,
+ "step": 2994
+ },
+ {
+ "epoch": 41.599303135888505,
+ "grad_norm": 0.04765695706009865,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 1.771855115890503,
+ "step": 2995
+ },
+ {
+ "epoch": 41.613240418118465,
+ "grad_norm": 0.04678108170628548,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 1.7572466135025024,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62717770034843,
+ "grad_norm": 0.04738544300198555,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 1.7798383235931396,
+ "step": 2997
+ },
+ {
+ "epoch": 41.641114982578394,
+ "grad_norm": 0.048733294010162354,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 1.7774200439453125,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65505226480836,
+ "grad_norm": 0.046759042888879776,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 1.777611255645752,
+ "step": 2999
+ },
+ {
+ "epoch": 41.66898954703833,
+ "grad_norm": 0.04731730744242668,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 1.7866392135620117,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68292682926829,
+ "grad_norm": 0.04734078422188759,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 1.7836531400680542,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69686411149826,
+ "grad_norm": 0.04705912247300148,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 1.7780985832214355,
+ "step": 3002
+ },
+ {
+ "epoch": 41.710801393728225,
+ "grad_norm": 0.048269402235746384,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 1.7863717079162598,
+ "step": 3003
+ },
+ {
+ "epoch": 41.724738675958186,
+ "grad_norm": 0.0475349985063076,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 1.7757699489593506,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73867595818815,
+ "grad_norm": 0.0482979416847229,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 1.7803300619125366,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75261324041812,
+ "grad_norm": 0.0486641526222229,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 1.7695112228393555,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76655052264808,
+ "grad_norm": 0.048009518533945084,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 1.7627280950546265,
+ "step": 3007
+ },
+ {
+ "epoch": 41.78048780487805,
+ "grad_norm": 0.04899517074227333,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 1.7776708602905273,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79442508710802,
+ "grad_norm": 0.047490231692790985,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 1.7935547828674316,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80836236933798,
+ "grad_norm": 0.04828681796789169,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 1.7704858779907227,
+ "step": 3010
+ },
+ {
+ "epoch": 41.822299651567945,
+ "grad_norm": 0.04789802059531212,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 1.7866098880767822,
+ "step": 3011
+ },
+ {
+ "epoch": 41.836236933797906,
+ "grad_norm": 0.04726841300725937,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 1.775240421295166,
+ "step": 3012
+ },
+ {
+ "epoch": 41.850174216027874,
+ "grad_norm": 0.048522237688302994,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 1.7788512706756592,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86411149825784,
+ "grad_norm": 0.04785652831196785,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 1.7695248126983643,
+ "step": 3014
+ },
+ {
+ "epoch": 41.8780487804878,
+ "grad_norm": 0.04859865456819534,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 1.768019437789917,
+ "step": 3015
+ },
+ {
+ "epoch": 41.89198606271777,
+ "grad_norm": 0.04948209971189499,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 1.7800569534301758,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90592334494774,
+ "grad_norm": 0.04760837182402611,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 1.7688634395599365,
+ "step": 3017
+ },
+ {
+ "epoch": 41.9198606271777,
+ "grad_norm": 0.04870005324482918,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 1.7903432846069336,
+ "step": 3018
+ },
+ {
+ "epoch": 41.933797909407666,
+ "grad_norm": 0.04837913066148758,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 1.7871191501617432,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94773519163763,
+ "grad_norm": 0.048451390117406845,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 1.7867250442504883,
+ "step": 3020
+ },
+ {
+ "epoch": 41.961672473867594,
+ "grad_norm": 0.04890143498778343,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 1.7735885381698608,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97560975609756,
+ "grad_norm": 0.04721386358141899,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 1.7692224979400635,
+ "step": 3022
+ },
+ {
+ "epoch": 41.98954703832753,
+ "grad_norm": 0.04814184084534645,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 1.776921033859253,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.039818305522203445,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 1.3309056758880615,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6534227132797241,
+ "eval_runtime": 40.6489,
+ "eval_samples_per_second": 60.075,
+ "eval_steps_per_second": 0.492,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01393728222997,
+ "grad_norm": 0.04341503232717514,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 1.7888715267181396,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02787456445993,
+ "grad_norm": 0.04266861081123352,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 1.772957682609558,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041811846689896,
+ "grad_norm": 0.043375164270401,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 1.7768115997314453,
+ "step": 3027
+ },
+ {
+ "epoch": 42.055749128919864,
+ "grad_norm": 0.044685594737529755,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 1.7511389255523682,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069686411149824,
+ "grad_norm": 0.0430758111178875,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 1.7582142353057861,
+ "step": 3029
+ },
+ {
+ "epoch": 42.08362369337979,
+ "grad_norm": 0.04426584020256996,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 1.7670388221740723,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09756097560975,
+ "grad_norm": 0.0441875085234642,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 1.7622346878051758,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11149825783972,
+ "grad_norm": 0.04547901079058647,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 1.7697657346725464,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12543554006969,
+ "grad_norm": 0.04314858466386795,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 1.7717976570129395,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13937282229965,
+ "grad_norm": 0.04366912320256233,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 1.7681739330291748,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153310104529616,
+ "grad_norm": 0.04399389028549194,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 1.763157844543457,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167247386759584,
+ "grad_norm": 0.043772801756858826,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 1.768390417098999,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181184668989545,
+ "grad_norm": 0.04431698098778725,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 1.7630479335784912,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19512195121951,
+ "grad_norm": 0.04417581111192703,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 1.771895408630371,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20905923344948,
+ "grad_norm": 0.045676663517951965,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 1.7609918117523193,
+ "step": 3039
+ },
+ {
+ "epoch": 42.22299651567944,
+ "grad_norm": 0.04374372959136963,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 1.7676066160202026,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23693379790941,
+ "grad_norm": 0.043877895921468735,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 1.7593367099761963,
+ "step": 3041
+ },
+ {
+ "epoch": 42.250871080139376,
+ "grad_norm": 0.0448957122862339,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 1.7552883625030518,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26480836236934,
+ "grad_norm": 0.04429576173424721,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 1.7635066509246826,
+ "step": 3043
+ },
+ {
+ "epoch": 42.278745644599304,
+ "grad_norm": 0.04319818317890167,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 1.7759575843811035,
+ "step": 3044
+ },
+ {
+ "epoch": 42.292682926829265,
+ "grad_norm": 0.04452989250421524,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 1.7780866622924805,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30662020905923,
+ "grad_norm": 0.044572584331035614,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 1.771972894668579,
+ "step": 3046
+ },
+ {
+ "epoch": 42.3205574912892,
+ "grad_norm": 0.044774096459150314,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 1.7678005695343018,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33449477351916,
+ "grad_norm": 0.0447823666036129,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 1.7687022686004639,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34843205574913,
+ "grad_norm": 0.04409811645746231,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 1.759227991104126,
+ "step": 3049
+ },
+ {
+ "epoch": 42.362369337979096,
+ "grad_norm": 0.04530395194888115,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 1.76583731174469,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37630662020906,
+ "grad_norm": 0.043990686535835266,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 1.776308536529541,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390243902439025,
+ "grad_norm": 0.045310501009225845,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 1.7690048217773438,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40418118466899,
+ "grad_norm": 0.045838840305805206,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 1.7719197273254395,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41811846689895,
+ "grad_norm": 0.044539984315633774,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 1.768972635269165,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43205574912892,
+ "grad_norm": 0.04546492174267769,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 1.7548890113830566,
+ "step": 3055
+ },
+ {
+ "epoch": 42.44599303135889,
+ "grad_norm": 0.0445762574672699,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 1.7676918506622314,
+ "step": 3056
+ },
+ {
+ "epoch": 42.45993031358885,
+ "grad_norm": 0.045404110103845596,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 1.7771313190460205,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47386759581882,
+ "grad_norm": 0.04481915012001991,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 1.7662057876586914,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48780487804878,
+ "grad_norm": 0.044174905866384506,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 1.7512820959091187,
+ "step": 3059
+ },
+ {
+ "epoch": 42.501742160278745,
+ "grad_norm": 0.04443437233567238,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 1.76678466796875,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51567944250871,
+ "grad_norm": 0.04414868354797363,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 1.7657546997070312,
+ "step": 3061
+ },
+ {
+ "epoch": 42.52961672473867,
+ "grad_norm": 0.045545708388090134,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 1.7497735023498535,
+ "step": 3062
+ },
+ {
+ "epoch": 42.54355400696864,
+ "grad_norm": 0.045831989496946335,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 1.7767338752746582,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55749128919861,
+ "grad_norm": 0.0455077588558197,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 1.7681231498718262,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57142857142857,
+ "grad_norm": 0.045490484684705734,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 1.7701823711395264,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58536585365854,
+ "grad_norm": 0.04617610201239586,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 1.7651724815368652,
+ "step": 3066
+ },
+ {
+ "epoch": 42.599303135888505,
+ "grad_norm": 0.045694541186094284,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 1.7682925462722778,
+ "step": 3067
+ },
+ {
+ "epoch": 42.613240418118465,
+ "grad_norm": 0.045455750077962875,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 1.771888017654419,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62717770034843,
+ "grad_norm": 0.04597267881035805,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 1.776789665222168,
+ "step": 3069
+ },
+ {
+ "epoch": 42.641114982578394,
+ "grad_norm": 0.046098366379737854,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 1.7781219482421875,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65505226480836,
+ "grad_norm": 0.04562405124306679,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 1.7662599086761475,
+ "step": 3071
+ },
+ {
+ "epoch": 42.66898954703833,
+ "grad_norm": 0.0461348257958889,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 1.775161623954773,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68292682926829,
+ "grad_norm": 0.04594029486179352,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 1.7788872718811035,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69686411149826,
+ "grad_norm": 0.0456400141119957,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 1.7744836807250977,
+ "step": 3074
+ },
+ {
+ "epoch": 42.710801393728225,
+ "grad_norm": 0.045577649027109146,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 1.7715539932250977,
+ "step": 3075
+ },
+ {
+ "epoch": 42.724738675958186,
+ "grad_norm": 0.046013571321964264,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 1.7707717418670654,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73867595818815,
+ "grad_norm": 0.046062543988227844,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 1.7822339534759521,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75261324041812,
+ "grad_norm": 0.04651106148958206,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 1.7592147588729858,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76655052264808,
+ "grad_norm": 0.046771664172410965,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 1.7719390392303467,
+ "step": 3079
+ },
+ {
+ "epoch": 42.78048780487805,
+ "grad_norm": 0.04678458347916603,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 1.7596237659454346,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79442508710802,
+ "grad_norm": 0.046154703944921494,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 1.7708245515823364,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80836236933798,
+ "grad_norm": 0.04633764550089836,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 1.7710938453674316,
+ "step": 3082
+ },
+ {
+ "epoch": 42.822299651567945,
+ "grad_norm": 0.046283479779958725,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 1.7665671110153198,
+ "step": 3083
+ },
+ {
+ "epoch": 42.836236933797906,
+ "grad_norm": 0.04648987576365471,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 1.7542476654052734,
+ "step": 3084
+ },
+ {
+ "epoch": 42.850174216027874,
+ "grad_norm": 0.04672298952937126,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 1.7689290046691895,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86411149825784,
+ "grad_norm": 0.046370647847652435,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 1.7649741172790527,
+ "step": 3086
+ },
+ {
+ "epoch": 42.8780487804878,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 1.7729518413543701,
+ "step": 3087
+ },
+ {
+ "epoch": 42.89198606271777,
+ "grad_norm": 0.04655609652400017,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 1.775625228881836,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90592334494774,
+ "grad_norm": 0.046239983290433884,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 1.7640939950942993,
+ "step": 3089
+ },
+ {
+ "epoch": 42.9198606271777,
+ "grad_norm": 0.046714168041944504,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 1.7750494480133057,
+ "step": 3090
+ },
+ {
+ "epoch": 42.933797909407666,
+ "grad_norm": 0.0479653924703598,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 1.7775356769561768,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94773519163763,
+ "grad_norm": 0.0469873771071434,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 1.7633222341537476,
+ "step": 3092
+ },
+ {
+ "epoch": 42.961672473867594,
+ "grad_norm": 0.04614265635609627,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 1.7667288780212402,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97560975609756,
+ "grad_norm": 0.04644224792718887,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 1.7725608348846436,
+ "step": 3094
+ },
+ {
+ "epoch": 42.98954703832753,
+ "grad_norm": 0.04625892639160156,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 1.7723143100738525,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03922572359442711,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 1.3379738330841064,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.654710054397583,
+ "eval_runtime": 40.5041,
+ "eval_samples_per_second": 60.29,
+ "eval_steps_per_second": 0.494,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01393728222997,
+ "grad_norm": 0.042021218687295914,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 1.7534828186035156,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02787456445993,
+ "grad_norm": 0.04172177240252495,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 1.7588303089141846,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041811846689896,
+ "grad_norm": 0.042629424482584,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 1.764153242111206,
+ "step": 3099
+ },
+ {
+ "epoch": 43.055749128919864,
+ "grad_norm": 0.04150637611746788,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 1.7654709815979004,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069686411149824,
+ "grad_norm": 0.042983464896678925,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 1.7520860433578491,
+ "step": 3101
+ },
+ {
+ "epoch": 43.08362369337979,
+ "grad_norm": 0.04234235733747482,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 1.7629724740982056,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09756097560975,
+ "grad_norm": 0.04364536330103874,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 1.7700273990631104,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11149825783972,
+ "grad_norm": 0.043138496577739716,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 1.7651735544204712,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12543554006969,
+ "grad_norm": 0.041396476328372955,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 1.7515389919281006,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13937282229965,
+ "grad_norm": 0.04298311099410057,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 1.7651766538619995,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153310104529616,
+ "grad_norm": 0.04275120422244072,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 1.7559534311294556,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167247386759584,
+ "grad_norm": 0.041940510272979736,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 1.7608134746551514,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181184668989545,
+ "grad_norm": 0.043498437851667404,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 1.7735629081726074,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19512195121951,
+ "grad_norm": 0.04355790466070175,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 1.7664821147918701,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20905923344948,
+ "grad_norm": 0.042053356766700745,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 1.7596155405044556,
+ "step": 3111
+ },
+ {
+ "epoch": 43.22299651567944,
+ "grad_norm": 0.04320033639669418,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 1.7779096364974976,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23693379790941,
+ "grad_norm": 0.042722251266241074,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 1.771970510482788,
+ "step": 3113
+ },
+ {
+ "epoch": 43.250871080139376,
+ "grad_norm": 0.043361593037843704,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 1.7600276470184326,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26480836236934,
+ "grad_norm": 0.04405055567622185,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 1.7721140384674072,
+ "step": 3115
+ },
+ {
+ "epoch": 43.278745644599304,
+ "grad_norm": 0.04226464778184891,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 1.755659818649292,
+ "step": 3116
+ },
+ {
+ "epoch": 43.292682926829265,
+ "grad_norm": 0.04350192844867706,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 1.7596216201782227,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30662020905923,
+ "grad_norm": 0.04462328553199768,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 1.7591629028320312,
+ "step": 3118
+ },
+ {
+ "epoch": 43.3205574912892,
+ "grad_norm": 0.04288632422685623,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 1.7450718879699707,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33449477351916,
+ "grad_norm": 0.0438922680914402,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 1.765263557434082,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34843205574913,
+ "grad_norm": 0.04307923838496208,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 1.7653264999389648,
+ "step": 3121
+ },
+ {
+ "epoch": 43.362369337979096,
+ "grad_norm": 0.04415101930499077,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 1.7691493034362793,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37630662020906,
+ "grad_norm": 0.04457402974367142,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 1.7685725688934326,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390243902439025,
+ "grad_norm": 0.04351523891091347,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 1.7489142417907715,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40418118466899,
+ "grad_norm": 0.043415915220975876,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 1.7557696104049683,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41811846689895,
+ "grad_norm": 0.044516436755657196,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 1.7655267715454102,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43205574912892,
+ "grad_norm": 0.04449203237891197,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 1.7631142139434814,
+ "step": 3127
+ },
+ {
+ "epoch": 43.44599303135889,
+ "grad_norm": 0.04321179911494255,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 1.7660303115844727,
+ "step": 3128
+ },
+ {
+ "epoch": 43.45993031358885,
+ "grad_norm": 0.04363849759101868,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 1.75204598903656,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47386759581882,
+ "grad_norm": 0.04379528760910034,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 1.7515747547149658,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48780487804878,
+ "grad_norm": 0.04358968138694763,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 1.7563517093658447,
+ "step": 3131
+ },
+ {
+ "epoch": 43.501742160278745,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 1.762972354888916,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51567944250871,
+ "grad_norm": 0.04360586032271385,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 1.7609463930130005,
+ "step": 3133
+ },
+ {
+ "epoch": 43.52961672473867,
+ "grad_norm": 0.04505139961838722,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 1.7723445892333984,
+ "step": 3134
+ },
+ {
+ "epoch": 43.54355400696864,
+ "grad_norm": 0.043562229722738266,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 1.7522692680358887,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55749128919861,
+ "grad_norm": 0.04364638775587082,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 1.7644182443618774,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57142857142857,
+ "grad_norm": 0.04494662210345268,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 1.7732231616973877,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58536585365854,
+ "grad_norm": 0.04464591667056084,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 1.754976749420166,
+ "step": 3138
+ },
+ {
+ "epoch": 43.599303135888505,
+ "grad_norm": 0.04465331509709358,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 1.7709685564041138,
+ "step": 3139
+ },
+ {
+ "epoch": 43.613240418118465,
+ "grad_norm": 0.04342262074351311,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 1.7601979970932007,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62717770034843,
+ "grad_norm": 0.04447892680764198,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 1.756757378578186,
+ "step": 3141
+ },
+ {
+ "epoch": 43.641114982578394,
+ "grad_norm": 0.04505394026637077,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 1.7673622369766235,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65505226480836,
+ "grad_norm": 0.044277530163526535,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 1.762864112854004,
+ "step": 3143
+ },
+ {
+ "epoch": 43.66898954703833,
+ "grad_norm": 0.04379425197839737,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 1.7678565979003906,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68292682926829,
+ "grad_norm": 0.04339452087879181,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 1.7612366676330566,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69686411149826,
+ "grad_norm": 0.04434647411108017,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 1.7668228149414062,
+ "step": 3146
+ },
+ {
+ "epoch": 43.710801393728225,
+ "grad_norm": 0.04466056823730469,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 1.7554584741592407,
+ "step": 3147
+ },
+ {
+ "epoch": 43.724738675958186,
+ "grad_norm": 0.044375862926244736,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 1.7701786756515503,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73867595818815,
+ "grad_norm": 0.044433314353227615,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 1.7586554288864136,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75261324041812,
+ "grad_norm": 0.0449117012321949,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 1.7642850875854492,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76655052264808,
+ "grad_norm": 0.04465561732649803,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 1.7654483318328857,
+ "step": 3151
+ },
+ {
+ "epoch": 43.78048780487805,
+ "grad_norm": 0.044917911291122437,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 1.7696160078048706,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79442508710802,
+ "grad_norm": 0.044988881796598434,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 1.7712256908416748,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80836236933798,
+ "grad_norm": 0.04482322186231613,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 1.762773871421814,
+ "step": 3154
+ },
+ {
+ "epoch": 43.822299651567945,
+ "grad_norm": 0.04423758387565613,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 1.7609460353851318,
+ "step": 3155
+ },
+ {
+ "epoch": 43.836236933797906,
+ "grad_norm": 0.044506512582302094,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 1.765155553817749,
+ "step": 3156
+ },
+ {
+ "epoch": 43.850174216027874,
+ "grad_norm": 0.04564986377954483,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 1.7608258724212646,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86411149825784,
+ "grad_norm": 0.043920136988162994,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 1.76609468460083,
+ "step": 3158
+ },
+ {
+ "epoch": 43.8780487804878,
+ "grad_norm": 0.04496336728334427,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 1.761902093887329,
+ "step": 3159
+ },
+ {
+ "epoch": 43.89198606271777,
+ "grad_norm": 0.0450543649494648,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 1.7570101022720337,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90592334494774,
+ "grad_norm": 0.04378526285290718,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 1.7584224939346313,
+ "step": 3161
+ },
+ {
+ "epoch": 43.9198606271777,
+ "grad_norm": 0.04425406455993652,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 1.7499665021896362,
+ "step": 3162
+ },
+ {
+ "epoch": 43.933797909407666,
+ "grad_norm": 0.044044118374586105,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 1.739232063293457,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94773519163763,
+ "grad_norm": 0.043724123388528824,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 1.7675073146820068,
+ "step": 3164
+ },
+ {
+ "epoch": 43.961672473867594,
+ "grad_norm": 0.04547949880361557,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 1.7621852159500122,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97560975609756,
+ "grad_norm": 0.04585639387369156,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 1.759611964225769,
+ "step": 3166
+ },
+ {
+ "epoch": 43.98954703832753,
+ "grad_norm": 0.04513043910264969,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 1.7627670764923096,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.037081386893987656,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 1.3273016214370728,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.655878484249115,
+ "eval_runtime": 62.6925,
+ "eval_samples_per_second": 38.952,
+ "eval_steps_per_second": 0.319,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01393728222997,
+ "grad_norm": 0.04096479341387749,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 1.7576605081558228,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02787456445993,
+ "grad_norm": 0.040468111634254456,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 1.7699885368347168,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041811846689896,
+ "grad_norm": 0.03982627019286156,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 1.741943597793579,
+ "step": 3171
+ },
+ {
+ "epoch": 44.055749128919864,
+ "grad_norm": 0.041303861886262894,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 1.7607184648513794,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069686411149824,
+ "grad_norm": 0.04077339917421341,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 1.7455109357833862,
+ "step": 3173
+ },
+ {
+ "epoch": 44.08362369337979,
+ "grad_norm": 0.04109608381986618,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 1.7644455432891846,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09756097560975,
+ "grad_norm": 0.04157571122050285,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 1.7602050304412842,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11149825783972,
+ "grad_norm": 0.040933798998594284,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 1.7585093975067139,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12543554006969,
+ "grad_norm": 0.040963515639305115,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 1.7687596082687378,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13937282229965,
+ "grad_norm": 0.042252764105796814,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 1.7694694995880127,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153310104529616,
+ "grad_norm": 0.04096998646855354,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 1.7650394439697266,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167247386759584,
+ "grad_norm": 0.04170529916882515,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 1.7502206563949585,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181184668989545,
+ "grad_norm": 0.04190446436405182,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 1.7572112083435059,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19512195121951,
+ "grad_norm": 0.042252860963344574,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 1.749457836151123,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20905923344948,
+ "grad_norm": 0.04197750985622406,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 1.7559109926223755,
+ "step": 3183
+ },
+ {
+ "epoch": 44.22299651567944,
+ "grad_norm": 0.04213518649339676,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 1.7385731935501099,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23693379790941,
+ "grad_norm": 0.04209589958190918,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 1.750757098197937,
+ "step": 3185
+ },
+ {
+ "epoch": 44.250871080139376,
+ "grad_norm": 0.04212355986237526,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 1.7661514282226562,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26480836236934,
+ "grad_norm": 0.042160581797361374,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 1.7580865621566772,
+ "step": 3187
+ },
+ {
+ "epoch": 44.278745644599304,
+ "grad_norm": 0.04218733310699463,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 1.7538390159606934,
+ "step": 3188
+ },
+ {
+ "epoch": 44.292682926829265,
+ "grad_norm": 0.04178664833307266,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 1.7542293071746826,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30662020905923,
+ "grad_norm": 0.04160580411553383,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 1.7468459606170654,
+ "step": 3190
+ },
+ {
+ "epoch": 44.3205574912892,
+ "grad_norm": 0.042538922280073166,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 1.7589120864868164,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33449477351916,
+ "grad_norm": 0.042292483150959015,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 1.759150743484497,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34843205574913,
+ "grad_norm": 0.04133346676826477,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 1.757230281829834,
+ "step": 3193
+ },
+ {
+ "epoch": 44.362369337979096,
+ "grad_norm": 0.042723916471004486,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 1.7720794677734375,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37630662020906,
+ "grad_norm": 0.04289544001221657,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 1.7658957242965698,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390243902439025,
+ "grad_norm": 0.0418098159134388,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 1.734413981437683,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40418118466899,
+ "grad_norm": 0.0412493497133255,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 1.7341864109039307,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41811846689895,
+ "grad_norm": 0.04258164390921593,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 1.7577321529388428,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43205574912892,
+ "grad_norm": 0.04071391373872757,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 1.7422387599945068,
+ "step": 3199
+ },
+ {
+ "epoch": 44.44599303135889,
+ "grad_norm": 0.04111524671316147,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 1.7563737630844116,
+ "step": 3200
+ },
+ {
+ "epoch": 44.45993031358885,
+ "grad_norm": 0.0422373004257679,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 1.7633488178253174,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47386759581882,
+ "grad_norm": 0.04157081991434097,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 1.7629499435424805,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48780487804878,
+ "grad_norm": 0.040883488953113556,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 1.7595467567443848,
+ "step": 3203
+ },
+ {
+ "epoch": 44.501742160278745,
+ "grad_norm": 0.04265756905078888,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 1.7728859186172485,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51567944250871,
+ "grad_norm": 0.04229922220110893,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 1.754500150680542,
+ "step": 3205
+ },
+ {
+ "epoch": 44.52961672473867,
+ "grad_norm": 0.04227421432733536,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 1.7667875289916992,
+ "step": 3206
+ },
+ {
+ "epoch": 44.54355400696864,
+ "grad_norm": 0.04271087795495987,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 1.7599055767059326,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55749128919861,
+ "grad_norm": 0.04237043857574463,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 1.74685537815094,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57142857142857,
+ "grad_norm": 0.042595189064741135,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 1.7581002712249756,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58536585365854,
+ "grad_norm": 0.04299062490463257,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 1.757879614830017,
+ "step": 3210
+ },
+ {
+ "epoch": 44.599303135888505,
+ "grad_norm": 0.04270675778388977,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 1.7636423110961914,
+ "step": 3211
+ },
+ {
+ "epoch": 44.613240418118465,
+ "grad_norm": 0.042374443262815475,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 1.7450188398361206,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62717770034843,
+ "grad_norm": 0.04340631887316704,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 1.7522510290145874,
+ "step": 3213
+ },
+ {
+ "epoch": 44.641114982578394,
+ "grad_norm": 0.04277212545275688,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 1.7763484716415405,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65505226480836,
+ "grad_norm": 0.04195157438516617,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 1.7633893489837646,
+ "step": 3215
+ },
+ {
+ "epoch": 44.66898954703833,
+ "grad_norm": 0.04262831434607506,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 1.7620947360992432,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68292682926829,
+ "grad_norm": 0.04257641360163689,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 1.757969856262207,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69686411149826,
+ "grad_norm": 0.04366584122180939,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 1.7606867551803589,
+ "step": 3218
+ },
+ {
+ "epoch": 44.710801393728225,
+ "grad_norm": 0.042488567531108856,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 1.750715732574463,
+ "step": 3219
+ },
+ {
+ "epoch": 44.724738675958186,
+ "grad_norm": 0.043666984885931015,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 1.756069302558899,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73867595818815,
+ "grad_norm": 0.04354606196284294,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 1.7556393146514893,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75261324041812,
+ "grad_norm": 0.042452357709407806,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 1.7572851181030273,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76655052264808,
+ "grad_norm": 0.043199196457862854,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 1.761836290359497,
+ "step": 3223
+ },
+ {
+ "epoch": 44.78048780487805,
+ "grad_norm": 0.04317186400294304,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 1.7503623962402344,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79442508710802,
+ "grad_norm": 0.04354375973343849,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 1.7610375881195068,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80836236933798,
+ "grad_norm": 0.042952246963977814,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 1.7500479221343994,
+ "step": 3226
+ },
+ {
+ "epoch": 44.822299651567945,
+ "grad_norm": 0.04257440194487572,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 1.7575607299804688,
+ "step": 3227
+ },
+ {
+ "epoch": 44.836236933797906,
+ "grad_norm": 0.041978977620601654,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 1.7537864446640015,
+ "step": 3228
+ },
+ {
+ "epoch": 44.850174216027874,
+ "grad_norm": 0.04331148788332939,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 1.772139072418213,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86411149825784,
+ "grad_norm": 0.04248177632689476,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 1.7485888004302979,
+ "step": 3230
+ },
+ {
+ "epoch": 44.8780487804878,
+ "grad_norm": 0.04298631474375725,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 1.7410145998001099,
+ "step": 3231
+ },
+ {
+ "epoch": 44.89198606271777,
+ "grad_norm": 0.0432223416864872,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 1.755599856376648,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90592334494774,
+ "grad_norm": 0.0427008792757988,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 1.764238715171814,
+ "step": 3233
+ },
+ {
+ "epoch": 44.9198606271777,
+ "grad_norm": 0.04341079294681549,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 1.7583544254302979,
+ "step": 3234
+ },
+ {
+ "epoch": 44.933797909407666,
+ "grad_norm": 0.0435490682721138,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 1.7442610263824463,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94773519163763,
+ "grad_norm": 0.042946282774209976,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 1.751840591430664,
+ "step": 3236
+ },
+ {
+ "epoch": 44.961672473867594,
+ "grad_norm": 0.0440446138381958,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 1.7394684553146362,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97560975609756,
+ "grad_norm": 0.04164465516805649,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 1.7417004108428955,
+ "step": 3238
+ },
+ {
+ "epoch": 44.98954703832753,
+ "grad_norm": 0.04329351335763931,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 1.7645809650421143,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.036814916878938675,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 1.3324573040008545,
+ "step": 3240
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.384945207392338e+18,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-3240/training_args.bin b/runs/mt5-ssl/checkpoint-3240/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3240/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-360/chat_template.jinja b/runs/mt5-ssl/checkpoint-360/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-360/config.json b/runs/mt5-ssl/checkpoint-360/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-360/generation_config.json b/runs/mt5-ssl/checkpoint-360/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-360/model.safetensors b/runs/mt5-ssl/checkpoint-360/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4d6afdba783de862011c7a5894d66a9ad34f189b
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7594160bf23c11aa3fd6224104b8bb7720579fce7fce24bb69c38e09cf578976
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-360/optimizer.pt b/runs/mt5-ssl/checkpoint-360/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..fb898788f7f61a5d0ff6bbc9bd705e99f7d69fcc
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3bf52918e3154a764ca1c8aacd3717a898cae36d7a5822b77382d86c81899816
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-360/rng_state_0.pth b/runs/mt5-ssl/checkpoint-360/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3a2bb760d0b0b50f4b7b6980626fda172a127f68
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9627a7b92eeed978894313d8642baf0ecdcd808f2102543d8b58136102ad5997
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-360/rng_state_1.pth b/runs/mt5-ssl/checkpoint-360/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7b307a1ab4859d8d838e7d20293e5eb15985549a
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0d79c621c2f337ae714092486c08d5a224dca6ddc4285fb0014e9b648d0c9cae
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-360/scheduler.pt b/runs/mt5-ssl/checkpoint-360/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..be1c46bbc9ac67b626f5e0bb3dac16ba107193bc
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:79ed78fca474f0ed71de3aa751471c35bef733680fc53c36641881688dd58ce5
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-360/tokenizer.json b/runs/mt5-ssl/checkpoint-360/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-360/tokenizer_config.json b/runs/mt5-ssl/checkpoint-360/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-360/trainer_state.json b/runs/mt5-ssl/checkpoint-360/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..2a0bc59e42f965f6e3a9fc18079881267375d71f
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/trainer_state.json
@@ -0,0 +1,2586 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 360,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5388280082137088e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-360/training_args.bin b/runs/mt5-ssl/checkpoint-360/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-360/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-3600/chat_template.jinja b/runs/mt5-ssl/checkpoint-3600/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-3600/config.json b/runs/mt5-ssl/checkpoint-3600/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-3600/generation_config.json b/runs/mt5-ssl/checkpoint-3600/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-3600/model.safetensors b/runs/mt5-ssl/checkpoint-3600/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ae8db4191b435101dfc6948e323a98eead1320c5
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9632e85cc7bcd531a0c4e41b246c07725e5061a0c40cf4eb1e035e0ce8812317
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-3600/optimizer.pt b/runs/mt5-ssl/checkpoint-3600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8fb6790f5d5ede54a09b952add9161489482c56d
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:baf1a3bc8def43bb8b0a9135ccc53263900c0885572f912ec7e860c5808cfc48
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-3600/rng_state_0.pth b/runs/mt5-ssl/checkpoint-3600/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..42e5fb760a4884c24f3899e0e2c34f9ab7d35d5e
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b9abac844de24b211d8b88d44e1866f089edee417850520a358afe34f866a821
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-3600/rng_state_1.pth b/runs/mt5-ssl/checkpoint-3600/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d6f10c7be65a25152763463abf33eb700d784288
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:68ba205d24a50bf8cdf4686014dfa7b5557eb04aefa63431dced1c3d75cbacab
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-3600/scheduler.pt b/runs/mt5-ssl/checkpoint-3600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..982d7b397b075cae944d790620cb0dfad6289952
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fcae2666f52aa083cc9b774f768ba87d551f0234d3e7800a3f2040f0f5905852
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-3600/tokenizer.json b/runs/mt5-ssl/checkpoint-3600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-3600/tokenizer_config.json b/runs/mt5-ssl/checkpoint-3600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-3600/trainer_state.json b/runs/mt5-ssl/checkpoint-3600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..418d5c5b18ea495074927e8564a0b97a7319d003
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/trainer_state.json
@@ -0,0 +1,25626 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 50.0,
+ "eval_steps": 500,
+ "global_step": 3600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 0.6166406273841858,
+ "eval_runtime": 39.9484,
+ "eval_samples_per_second": 61.129,
+ "eval_steps_per_second": 0.501,
+ "step": 720
+ },
+ {
+ "epoch": 10.013937282229966,
+ "grad_norm": 0.03875136002898216,
+ "learning_rate": 0.0005551837152119915,
+ "loss": 2.3220555782318115,
+ "step": 721
+ },
+ {
+ "epoch": 10.02787456445993,
+ "grad_norm": 0.056541845202445984,
+ "learning_rate": 0.0005550427955088455,
+ "loss": 2.3221559524536133,
+ "step": 722
+ },
+ {
+ "epoch": 10.041811846689896,
+ "grad_norm": 0.06505046039819717,
+ "learning_rate": 0.0005549016727393715,
+ "loss": 2.3312087059020996,
+ "step": 723
+ },
+ {
+ "epoch": 10.05574912891986,
+ "grad_norm": 0.066843181848526,
+ "learning_rate": 0.0005547603470172697,
+ "loss": 2.3008155822753906,
+ "step": 724
+ },
+ {
+ "epoch": 10.069686411149826,
+ "grad_norm": 0.07698776572942734,
+ "learning_rate": 0.0005546188184564036,
+ "loss": 2.3316988945007324,
+ "step": 725
+ },
+ {
+ "epoch": 10.08362369337979,
+ "grad_norm": 0.08154117316007614,
+ "learning_rate": 0.0005544770871708004,
+ "loss": 2.332336902618408,
+ "step": 726
+ },
+ {
+ "epoch": 10.097560975609756,
+ "grad_norm": 0.07038415223360062,
+ "learning_rate": 0.0005543351532746501,
+ "loss": 2.3396337032318115,
+ "step": 727
+ },
+ {
+ "epoch": 10.111498257839722,
+ "grad_norm": 0.07695100456476212,
+ "learning_rate": 0.0005541930168823067,
+ "loss": 2.321103811264038,
+ "step": 728
+ },
+ {
+ "epoch": 10.125435540069686,
+ "grad_norm": 0.09513774514198303,
+ "learning_rate": 0.0005540506781082864,
+ "loss": 2.341754913330078,
+ "step": 729
+ },
+ {
+ "epoch": 10.139372822299652,
+ "grad_norm": 0.10298766195774078,
+ "learning_rate": 0.0005539081370672695,
+ "loss": 2.326603412628174,
+ "step": 730
+ },
+ {
+ "epoch": 10.153310104529616,
+ "grad_norm": 0.08832718431949615,
+ "learning_rate": 0.0005537653938740984,
+ "loss": 2.3495705127716064,
+ "step": 731
+ },
+ {
+ "epoch": 10.167247386759582,
+ "grad_norm": 0.08626274764537811,
+ "learning_rate": 0.0005536224486437788,
+ "loss": 2.3341174125671387,
+ "step": 732
+ },
+ {
+ "epoch": 10.181184668989546,
+ "grad_norm": 0.09540148824453354,
+ "learning_rate": 0.0005534793014914793,
+ "loss": 2.3632583618164062,
+ "step": 733
+ },
+ {
+ "epoch": 10.195121951219512,
+ "grad_norm": 0.09003570675849915,
+ "learning_rate": 0.0005533359525325307,
+ "loss": 2.359584093093872,
+ "step": 734
+ },
+ {
+ "epoch": 10.209059233449477,
+ "grad_norm": 0.08710908889770508,
+ "learning_rate": 0.0005531924018824266,
+ "loss": 2.358776330947876,
+ "step": 735
+ },
+ {
+ "epoch": 10.222996515679442,
+ "grad_norm": 0.09109727293252945,
+ "learning_rate": 0.0005530486496568236,
+ "loss": 2.3650546073913574,
+ "step": 736
+ },
+ {
+ "epoch": 10.236933797909408,
+ "grad_norm": 0.09278365224599838,
+ "learning_rate": 0.0005529046959715397,
+ "loss": 2.355679988861084,
+ "step": 737
+ },
+ {
+ "epoch": 10.250871080139373,
+ "grad_norm": 0.08022253215312958,
+ "learning_rate": 0.0005527605409425562,
+ "loss": 2.3479089736938477,
+ "step": 738
+ },
+ {
+ "epoch": 10.264808362369338,
+ "grad_norm": 0.0726609081029892,
+ "learning_rate": 0.0005526161846860161,
+ "loss": 2.343989133834839,
+ "step": 739
+ },
+ {
+ "epoch": 10.278745644599303,
+ "grad_norm": 0.06493718922138214,
+ "learning_rate": 0.0005524716273182245,
+ "loss": 2.3520140647888184,
+ "step": 740
+ },
+ {
+ "epoch": 10.292682926829269,
+ "grad_norm": 0.05545816943049431,
+ "learning_rate": 0.0005523268689556489,
+ "loss": 2.326129198074341,
+ "step": 741
+ },
+ {
+ "epoch": 10.306620209059233,
+ "grad_norm": 0.054707564413547516,
+ "learning_rate": 0.0005521819097149183,
+ "loss": 2.3311123847961426,
+ "step": 742
+ },
+ {
+ "epoch": 10.320557491289199,
+ "grad_norm": 0.05621136724948883,
+ "learning_rate": 0.0005520367497128238,
+ "loss": 2.3465726375579834,
+ "step": 743
+ },
+ {
+ "epoch": 10.334494773519165,
+ "grad_norm": 0.053385693579912186,
+ "learning_rate": 0.0005518913890663182,
+ "loss": 2.366509437561035,
+ "step": 744
+ },
+ {
+ "epoch": 10.348432055749129,
+ "grad_norm": 0.05938078835606575,
+ "learning_rate": 0.0005517458278925161,
+ "loss": 2.35367488861084,
+ "step": 745
+ },
+ {
+ "epoch": 10.362369337979095,
+ "grad_norm": 0.052808232605457306,
+ "learning_rate": 0.0005516000663086932,
+ "loss": 2.3501806259155273,
+ "step": 746
+ },
+ {
+ "epoch": 10.376306620209059,
+ "grad_norm": 0.04525364562869072,
+ "learning_rate": 0.0005514541044322872,
+ "loss": 2.3461456298828125,
+ "step": 747
+ },
+ {
+ "epoch": 10.390243902439025,
+ "grad_norm": 0.04237867519259453,
+ "learning_rate": 0.0005513079423808967,
+ "loss": 2.354640483856201,
+ "step": 748
+ },
+ {
+ "epoch": 10.404181184668989,
+ "grad_norm": 0.041736457496881485,
+ "learning_rate": 0.0005511615802722823,
+ "loss": 2.348954916000366,
+ "step": 749
+ },
+ {
+ "epoch": 10.418118466898955,
+ "grad_norm": 0.03999985754489899,
+ "learning_rate": 0.0005510150182243647,
+ "loss": 2.3511078357696533,
+ "step": 750
+ },
+ {
+ "epoch": 10.43205574912892,
+ "grad_norm": 0.03754125162959099,
+ "learning_rate": 0.0005508682563552269,
+ "loss": 2.3245179653167725,
+ "step": 751
+ },
+ {
+ "epoch": 10.445993031358885,
+ "grad_norm": 0.03457164764404297,
+ "learning_rate": 0.0005507212947831119,
+ "loss": 2.3434243202209473,
+ "step": 752
+ },
+ {
+ "epoch": 10.45993031358885,
+ "grad_norm": 0.03329804539680481,
+ "learning_rate": 0.0005505741336264239,
+ "loss": 2.338502883911133,
+ "step": 753
+ },
+ {
+ "epoch": 10.473867595818815,
+ "grad_norm": 0.03289134055376053,
+ "learning_rate": 0.0005504267730037283,
+ "loss": 2.325770378112793,
+ "step": 754
+ },
+ {
+ "epoch": 10.487804878048781,
+ "grad_norm": 0.03277945891022682,
+ "learning_rate": 0.0005502792130337508,
+ "loss": 2.3254008293151855,
+ "step": 755
+ },
+ {
+ "epoch": 10.501742160278745,
+ "grad_norm": 0.03316863626241684,
+ "learning_rate": 0.0005501314538353776,
+ "loss": 2.3330087661743164,
+ "step": 756
+ },
+ {
+ "epoch": 10.515679442508711,
+ "grad_norm": 0.032691389322280884,
+ "learning_rate": 0.000549983495527656,
+ "loss": 2.3320021629333496,
+ "step": 757
+ },
+ {
+ "epoch": 10.529616724738675,
+ "grad_norm": 0.03377344831824303,
+ "learning_rate": 0.000549835338229793,
+ "loss": 2.3412442207336426,
+ "step": 758
+ },
+ {
+ "epoch": 10.543554006968641,
+ "grad_norm": 0.0311700738966465,
+ "learning_rate": 0.0005496869820611563,
+ "loss": 2.3270277976989746,
+ "step": 759
+ },
+ {
+ "epoch": 10.557491289198607,
+ "grad_norm": 0.03159717842936516,
+ "learning_rate": 0.0005495384271412739,
+ "loss": 2.332291603088379,
+ "step": 760
+ },
+ {
+ "epoch": 10.571428571428571,
+ "grad_norm": 0.03267215937376022,
+ "learning_rate": 0.0005493896735898338,
+ "loss": 2.3158648014068604,
+ "step": 761
+ },
+ {
+ "epoch": 10.585365853658537,
+ "grad_norm": 0.03247901797294617,
+ "learning_rate": 0.000549240721526684,
+ "loss": 2.3411662578582764,
+ "step": 762
+ },
+ {
+ "epoch": 10.599303135888501,
+ "grad_norm": 0.033130206167697906,
+ "learning_rate": 0.0005490915710718325,
+ "loss": 2.3163256645202637,
+ "step": 763
+ },
+ {
+ "epoch": 10.613240418118467,
+ "grad_norm": 0.03357957676053047,
+ "learning_rate": 0.0005489422223454471,
+ "loss": 2.3357491493225098,
+ "step": 764
+ },
+ {
+ "epoch": 10.627177700348431,
+ "grad_norm": 0.033023931086063385,
+ "learning_rate": 0.0005487926754678556,
+ "loss": 2.3287644386291504,
+ "step": 765
+ },
+ {
+ "epoch": 10.641114982578397,
+ "grad_norm": 0.031831514090299606,
+ "learning_rate": 0.0005486429305595448,
+ "loss": 2.3300392627716064,
+ "step": 766
+ },
+ {
+ "epoch": 10.655052264808361,
+ "grad_norm": 0.033411115407943726,
+ "learning_rate": 0.0005484929877411618,
+ "loss": 2.3265159130096436,
+ "step": 767
+ },
+ {
+ "epoch": 10.668989547038327,
+ "grad_norm": 0.03543822839856148,
+ "learning_rate": 0.0005483428471335127,
+ "loss": 2.3355298042297363,
+ "step": 768
+ },
+ {
+ "epoch": 10.682926829268293,
+ "grad_norm": 0.035002339631319046,
+ "learning_rate": 0.000548192508857563,
+ "loss": 2.339534282684326,
+ "step": 769
+ },
+ {
+ "epoch": 10.696864111498257,
+ "grad_norm": 0.03508792817592621,
+ "learning_rate": 0.0005480419730344377,
+ "loss": 2.3224260807037354,
+ "step": 770
+ },
+ {
+ "epoch": 10.710801393728223,
+ "grad_norm": 0.03275252878665924,
+ "learning_rate": 0.0005478912397854208,
+ "loss": 2.3294410705566406,
+ "step": 771
+ },
+ {
+ "epoch": 10.724738675958188,
+ "grad_norm": 0.03358299657702446,
+ "learning_rate": 0.0005477403092319552,
+ "loss": 2.3402085304260254,
+ "step": 772
+ },
+ {
+ "epoch": 10.738675958188153,
+ "grad_norm": 0.03494272753596306,
+ "learning_rate": 0.0005475891814956429,
+ "loss": 2.3288674354553223,
+ "step": 773
+ },
+ {
+ "epoch": 10.752613240418118,
+ "grad_norm": 0.034237124025821686,
+ "learning_rate": 0.0005474378566982449,
+ "loss": 2.3190884590148926,
+ "step": 774
+ },
+ {
+ "epoch": 10.766550522648084,
+ "grad_norm": 0.03367462381720543,
+ "learning_rate": 0.0005472863349616809,
+ "loss": 2.330655097961426,
+ "step": 775
+ },
+ {
+ "epoch": 10.78048780487805,
+ "grad_norm": 0.03104310855269432,
+ "learning_rate": 0.0005471346164080292,
+ "loss": 2.326869487762451,
+ "step": 776
+ },
+ {
+ "epoch": 10.794425087108014,
+ "grad_norm": 0.033606868237257004,
+ "learning_rate": 0.0005469827011595266,
+ "loss": 2.3256993293762207,
+ "step": 777
+ },
+ {
+ "epoch": 10.80836236933798,
+ "grad_norm": 0.030802417546510696,
+ "learning_rate": 0.0005468305893385686,
+ "loss": 2.3224363327026367,
+ "step": 778
+ },
+ {
+ "epoch": 10.822299651567944,
+ "grad_norm": 0.03256989270448685,
+ "learning_rate": 0.0005466782810677088,
+ "loss": 2.3356478214263916,
+ "step": 779
+ },
+ {
+ "epoch": 10.83623693379791,
+ "grad_norm": 0.03525761514902115,
+ "learning_rate": 0.0005465257764696593,
+ "loss": 2.323814868927002,
+ "step": 780
+ },
+ {
+ "epoch": 10.850174216027874,
+ "grad_norm": 0.03283970057964325,
+ "learning_rate": 0.0005463730756672903,
+ "loss": 2.328904628753662,
+ "step": 781
+ },
+ {
+ "epoch": 10.86411149825784,
+ "grad_norm": 0.030748914927244186,
+ "learning_rate": 0.0005462201787836301,
+ "loss": 2.3218486309051514,
+ "step": 782
+ },
+ {
+ "epoch": 10.878048780487806,
+ "grad_norm": 0.03220753371715546,
+ "learning_rate": 0.0005460670859418651,
+ "loss": 2.3403337001800537,
+ "step": 783
+ },
+ {
+ "epoch": 10.89198606271777,
+ "grad_norm": 0.033538568764925,
+ "learning_rate": 0.000545913797265339,
+ "loss": 2.347982406616211,
+ "step": 784
+ },
+ {
+ "epoch": 10.905923344947736,
+ "grad_norm": 0.030491137877106667,
+ "learning_rate": 0.0005457603128775541,
+ "loss": 2.330781936645508,
+ "step": 785
+ },
+ {
+ "epoch": 10.9198606271777,
+ "grad_norm": 0.03357788547873497,
+ "learning_rate": 0.0005456066329021699,
+ "loss": 2.3255670070648193,
+ "step": 786
+ },
+ {
+ "epoch": 10.933797909407666,
+ "grad_norm": 0.033416833728551865,
+ "learning_rate": 0.0005454527574630037,
+ "loss": 2.324746608734131,
+ "step": 787
+ },
+ {
+ "epoch": 10.94773519163763,
+ "grad_norm": 0.030557911843061447,
+ "learning_rate": 0.0005452986866840298,
+ "loss": 2.3319931030273438,
+ "step": 788
+ },
+ {
+ "epoch": 10.961672473867596,
+ "grad_norm": 0.03264349699020386,
+ "learning_rate": 0.0005451444206893806,
+ "loss": 2.325213670730591,
+ "step": 789
+ },
+ {
+ "epoch": 10.975609756097562,
+ "grad_norm": 0.03614091873168945,
+ "learning_rate": 0.0005449899596033453,
+ "loss": 2.3426029682159424,
+ "step": 790
+ },
+ {
+ "epoch": 10.989547038327526,
+ "grad_norm": 0.03389138728380203,
+ "learning_rate": 0.0005448353035503704,
+ "loss": 2.3327999114990234,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.027671724557876587,
+ "learning_rate": 0.0005446804526550596,
+ "loss": 1.7424343824386597,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 0.6149941086769104,
+ "eval_runtime": 40.6376,
+ "eval_samples_per_second": 60.092,
+ "eval_steps_per_second": 0.492,
+ "step": 792
+ },
+ {
+ "epoch": 11.013937282229966,
+ "grad_norm": 0.041478194296360016,
+ "learning_rate": 0.0005445254070421732,
+ "loss": 2.2832956314086914,
+ "step": 793
+ },
+ {
+ "epoch": 11.02787456445993,
+ "grad_norm": 0.052794840186834335,
+ "learning_rate": 0.0005443701668366288,
+ "loss": 2.275270462036133,
+ "step": 794
+ },
+ {
+ "epoch": 11.041811846689896,
+ "grad_norm": 0.0529312826693058,
+ "learning_rate": 0.0005442147321635007,
+ "loss": 2.2793469429016113,
+ "step": 795
+ },
+ {
+ "epoch": 11.05574912891986,
+ "grad_norm": 0.053920548409223557,
+ "learning_rate": 0.00054405910314802,
+ "loss": 2.282958507537842,
+ "step": 796
+ },
+ {
+ "epoch": 11.069686411149826,
+ "grad_norm": 0.053466152399778366,
+ "learning_rate": 0.0005439032799155737,
+ "loss": 2.291733503341675,
+ "step": 797
+ },
+ {
+ "epoch": 11.08362369337979,
+ "grad_norm": 0.06293711066246033,
+ "learning_rate": 0.000543747262591706,
+ "loss": 2.286085605621338,
+ "step": 798
+ },
+ {
+ "epoch": 11.097560975609756,
+ "grad_norm": 0.07056494057178497,
+ "learning_rate": 0.0005435910513021172,
+ "loss": 2.291978359222412,
+ "step": 799
+ },
+ {
+ "epoch": 11.111498257839722,
+ "grad_norm": 0.06913711130619049,
+ "learning_rate": 0.0005434346461726641,
+ "loss": 2.286970853805542,
+ "step": 800
+ },
+ {
+ "epoch": 11.125435540069686,
+ "grad_norm": 0.07270195335149765,
+ "learning_rate": 0.0005432780473293593,
+ "loss": 2.293346643447876,
+ "step": 801
+ },
+ {
+ "epoch": 11.139372822299652,
+ "grad_norm": 0.07776486128568649,
+ "learning_rate": 0.0005431212548983718,
+ "loss": 2.2992706298828125,
+ "step": 802
+ },
+ {
+ "epoch": 11.153310104529616,
+ "grad_norm": 0.0768241286277771,
+ "learning_rate": 0.0005429642690060262,
+ "loss": 2.292184352874756,
+ "step": 803
+ },
+ {
+ "epoch": 11.167247386759582,
+ "grad_norm": 0.07126478105783463,
+ "learning_rate": 0.0005428070897788034,
+ "loss": 2.2916717529296875,
+ "step": 804
+ },
+ {
+ "epoch": 11.181184668989546,
+ "grad_norm": 0.05803872272372246,
+ "learning_rate": 0.0005426497173433398,
+ "loss": 2.3002567291259766,
+ "step": 805
+ },
+ {
+ "epoch": 11.195121951219512,
+ "grad_norm": 0.062250033020973206,
+ "learning_rate": 0.0005424921518264275,
+ "loss": 2.280787229537964,
+ "step": 806
+ },
+ {
+ "epoch": 11.209059233449477,
+ "grad_norm": 0.06336672604084015,
+ "learning_rate": 0.0005423343933550143,
+ "loss": 2.299450397491455,
+ "step": 807
+ },
+ {
+ "epoch": 11.222996515679442,
+ "grad_norm": 0.061671603471040726,
+ "learning_rate": 0.0005421764420562032,
+ "loss": 2.299621105194092,
+ "step": 808
+ },
+ {
+ "epoch": 11.236933797909408,
+ "grad_norm": 0.050586555153131485,
+ "learning_rate": 0.0005420182980572527,
+ "loss": 2.2990193367004395,
+ "step": 809
+ },
+ {
+ "epoch": 11.250871080139373,
+ "grad_norm": 0.046991016715765,
+ "learning_rate": 0.0005418599614855768,
+ "loss": 2.2989754676818848,
+ "step": 810
+ },
+ {
+ "epoch": 11.264808362369338,
+ "grad_norm": 0.049337320029735565,
+ "learning_rate": 0.0005417014324687442,
+ "loss": 2.3104770183563232,
+ "step": 811
+ },
+ {
+ "epoch": 11.278745644599303,
+ "grad_norm": 0.04728561267256737,
+ "learning_rate": 0.000541542711134479,
+ "loss": 2.294708251953125,
+ "step": 812
+ },
+ {
+ "epoch": 11.292682926829269,
+ "grad_norm": 0.04332006350159645,
+ "learning_rate": 0.0005413837976106599,
+ "loss": 2.2984559535980225,
+ "step": 813
+ },
+ {
+ "epoch": 11.306620209059233,
+ "grad_norm": 0.04177679494023323,
+ "learning_rate": 0.0005412246920253208,
+ "loss": 2.2913899421691895,
+ "step": 814
+ },
+ {
+ "epoch": 11.320557491289199,
+ "grad_norm": 0.04118878394365311,
+ "learning_rate": 0.0005410653945066503,
+ "loss": 2.2980523109436035,
+ "step": 815
+ },
+ {
+ "epoch": 11.334494773519165,
+ "grad_norm": 0.03696153685450554,
+ "learning_rate": 0.0005409059051829914,
+ "loss": 2.291243314743042,
+ "step": 816
+ },
+ {
+ "epoch": 11.348432055749129,
+ "grad_norm": 0.03917296603322029,
+ "learning_rate": 0.0005407462241828417,
+ "loss": 2.291288375854492,
+ "step": 817
+ },
+ {
+ "epoch": 11.362369337979095,
+ "grad_norm": 0.0412733256816864,
+ "learning_rate": 0.0005405863516348535,
+ "loss": 2.3113622665405273,
+ "step": 818
+ },
+ {
+ "epoch": 11.376306620209059,
+ "grad_norm": 0.047993581742048264,
+ "learning_rate": 0.000540426287667833,
+ "loss": 2.3065576553344727,
+ "step": 819
+ },
+ {
+ "epoch": 11.390243902439025,
+ "grad_norm": 0.04606267437338829,
+ "learning_rate": 0.000540266032410741,
+ "loss": 2.2904818058013916,
+ "step": 820
+ },
+ {
+ "epoch": 11.404181184668989,
+ "grad_norm": 0.04308634623885155,
+ "learning_rate": 0.0005401055859926923,
+ "loss": 2.30252742767334,
+ "step": 821
+ },
+ {
+ "epoch": 11.418118466898955,
+ "grad_norm": 0.04426122456789017,
+ "learning_rate": 0.0005399449485429555,
+ "loss": 2.300410747528076,
+ "step": 822
+ },
+ {
+ "epoch": 11.43205574912892,
+ "grad_norm": 0.04351365193724632,
+ "learning_rate": 0.0005397841201909535,
+ "loss": 2.3004343509674072,
+ "step": 823
+ },
+ {
+ "epoch": 11.445993031358885,
+ "grad_norm": 0.04511446878314018,
+ "learning_rate": 0.0005396231010662627,
+ "loss": 2.2854676246643066,
+ "step": 824
+ },
+ {
+ "epoch": 11.45993031358885,
+ "grad_norm": 0.044850338250398636,
+ "learning_rate": 0.0005394618912986132,
+ "loss": 2.295956611633301,
+ "step": 825
+ },
+ {
+ "epoch": 11.473867595818815,
+ "grad_norm": 0.0453290119767189,
+ "learning_rate": 0.0005393004910178889,
+ "loss": 2.31127667427063,
+ "step": 826
+ },
+ {
+ "epoch": 11.487804878048781,
+ "grad_norm": 0.045144420117139816,
+ "learning_rate": 0.0005391389003541271,
+ "loss": 2.2946109771728516,
+ "step": 827
+ },
+ {
+ "epoch": 11.501742160278745,
+ "grad_norm": 0.04222119599580765,
+ "learning_rate": 0.0005389771194375185,
+ "loss": 2.3143534660339355,
+ "step": 828
+ },
+ {
+ "epoch": 11.515679442508711,
+ "grad_norm": 0.0382888987660408,
+ "learning_rate": 0.0005388151483984071,
+ "loss": 2.2896716594696045,
+ "step": 829
+ },
+ {
+ "epoch": 11.529616724738675,
+ "grad_norm": 0.04116874188184738,
+ "learning_rate": 0.0005386529873672898,
+ "loss": 2.2939164638519287,
+ "step": 830
+ },
+ {
+ "epoch": 11.543554006968641,
+ "grad_norm": 0.0424816831946373,
+ "learning_rate": 0.0005384906364748172,
+ "loss": 2.286743402481079,
+ "step": 831
+ },
+ {
+ "epoch": 11.557491289198607,
+ "grad_norm": 0.04002925008535385,
+ "learning_rate": 0.000538328095851792,
+ "loss": 2.308349847793579,
+ "step": 832
+ },
+ {
+ "epoch": 11.571428571428571,
+ "grad_norm": 0.04064339026808739,
+ "learning_rate": 0.0005381653656291707,
+ "loss": 2.2800822257995605,
+ "step": 833
+ },
+ {
+ "epoch": 11.585365853658537,
+ "grad_norm": 0.04154384508728981,
+ "learning_rate": 0.0005380024459380619,
+ "loss": 2.286264657974243,
+ "step": 834
+ },
+ {
+ "epoch": 11.599303135888501,
+ "grad_norm": 0.039285432547330856,
+ "learning_rate": 0.000537839336909727,
+ "loss": 2.308560848236084,
+ "step": 835
+ },
+ {
+ "epoch": 11.613240418118467,
+ "grad_norm": 0.0349508672952652,
+ "learning_rate": 0.00053767603867558,
+ "loss": 2.318333625793457,
+ "step": 836
+ },
+ {
+ "epoch": 11.627177700348431,
+ "grad_norm": 0.03986033424735069,
+ "learning_rate": 0.0005375125513671874,
+ "loss": 2.2920660972595215,
+ "step": 837
+ },
+ {
+ "epoch": 11.641114982578397,
+ "grad_norm": 0.039767954498529434,
+ "learning_rate": 0.0005373488751162681,
+ "loss": 2.305238962173462,
+ "step": 838
+ },
+ {
+ "epoch": 11.655052264808361,
+ "grad_norm": 0.036252446472644806,
+ "learning_rate": 0.0005371850100546929,
+ "loss": 2.297325611114502,
+ "step": 839
+ },
+ {
+ "epoch": 11.668989547038327,
+ "grad_norm": 0.03562362864613533,
+ "learning_rate": 0.000537020956314485,
+ "loss": 2.2985639572143555,
+ "step": 840
+ },
+ {
+ "epoch": 11.682926829268293,
+ "grad_norm": 0.04167954996228218,
+ "learning_rate": 0.0005368567140278196,
+ "loss": 2.277585506439209,
+ "step": 841
+ },
+ {
+ "epoch": 11.696864111498257,
+ "grad_norm": 0.04170938953757286,
+ "learning_rate": 0.0005366922833270235,
+ "loss": 2.301539897918701,
+ "step": 842
+ },
+ {
+ "epoch": 11.710801393728223,
+ "grad_norm": 0.03688978776335716,
+ "learning_rate": 0.0005365276643445757,
+ "loss": 2.3004941940307617,
+ "step": 843
+ },
+ {
+ "epoch": 11.724738675958188,
+ "grad_norm": 0.03725716844201088,
+ "learning_rate": 0.0005363628572131069,
+ "loss": 2.2922308444976807,
+ "step": 844
+ },
+ {
+ "epoch": 11.738675958188153,
+ "grad_norm": 0.038027532398700714,
+ "learning_rate": 0.0005361978620653991,
+ "loss": 2.301987648010254,
+ "step": 845
+ },
+ {
+ "epoch": 11.752613240418118,
+ "grad_norm": 0.03414970636367798,
+ "learning_rate": 0.0005360326790343858,
+ "loss": 2.293957233428955,
+ "step": 846
+ },
+ {
+ "epoch": 11.766550522648084,
+ "grad_norm": 0.03756559267640114,
+ "learning_rate": 0.000535867308253152,
+ "loss": 2.3135273456573486,
+ "step": 847
+ },
+ {
+ "epoch": 11.78048780487805,
+ "grad_norm": 0.035104189068078995,
+ "learning_rate": 0.0005357017498549341,
+ "loss": 2.3000903129577637,
+ "step": 848
+ },
+ {
+ "epoch": 11.794425087108014,
+ "grad_norm": 0.0360398069024086,
+ "learning_rate": 0.0005355360039731196,
+ "loss": 2.300976276397705,
+ "step": 849
+ },
+ {
+ "epoch": 11.80836236933798,
+ "grad_norm": 0.037384308874607086,
+ "learning_rate": 0.0005353700707412466,
+ "loss": 2.295656681060791,
+ "step": 850
+ },
+ {
+ "epoch": 11.822299651567944,
+ "grad_norm": 0.03750584274530411,
+ "learning_rate": 0.0005352039502930048,
+ "loss": 2.3190412521362305,
+ "step": 851
+ },
+ {
+ "epoch": 11.83623693379791,
+ "grad_norm": 0.038159798830747604,
+ "learning_rate": 0.0005350376427622343,
+ "loss": 2.302290916442871,
+ "step": 852
+ },
+ {
+ "epoch": 11.850174216027874,
+ "grad_norm": 0.03608061745762825,
+ "learning_rate": 0.0005348711482829261,
+ "loss": 2.3079752922058105,
+ "step": 853
+ },
+ {
+ "epoch": 11.86411149825784,
+ "grad_norm": 0.0350031703710556,
+ "learning_rate": 0.000534704466989222,
+ "loss": 2.303936004638672,
+ "step": 854
+ },
+ {
+ "epoch": 11.878048780487806,
+ "grad_norm": 0.03555091843008995,
+ "learning_rate": 0.0005345375990154138,
+ "loss": 2.2996826171875,
+ "step": 855
+ },
+ {
+ "epoch": 11.89198606271777,
+ "grad_norm": 0.03820710629224777,
+ "learning_rate": 0.0005343705444959441,
+ "loss": 2.3136544227600098,
+ "step": 856
+ },
+ {
+ "epoch": 11.905923344947736,
+ "grad_norm": 0.0389101468026638,
+ "learning_rate": 0.000534203303565406,
+ "loss": 2.2767484188079834,
+ "step": 857
+ },
+ {
+ "epoch": 11.9198606271777,
+ "grad_norm": 0.03647430241107941,
+ "learning_rate": 0.000534035876358542,
+ "loss": 2.2977707386016846,
+ "step": 858
+ },
+ {
+ "epoch": 11.933797909407666,
+ "grad_norm": 0.038748398423194885,
+ "learning_rate": 0.0005338682630102454,
+ "loss": 2.2904186248779297,
+ "step": 859
+ },
+ {
+ "epoch": 11.94773519163763,
+ "grad_norm": 0.03643745929002762,
+ "learning_rate": 0.0005337004636555595,
+ "loss": 2.296473503112793,
+ "step": 860
+ },
+ {
+ "epoch": 11.961672473867596,
+ "grad_norm": 0.03438796103000641,
+ "learning_rate": 0.0005335324784296768,
+ "loss": 2.309252977371216,
+ "step": 861
+ },
+ {
+ "epoch": 11.975609756097562,
+ "grad_norm": 0.0365566685795784,
+ "learning_rate": 0.0005333643074679403,
+ "loss": 2.310616970062256,
+ "step": 862
+ },
+ {
+ "epoch": 11.989547038327526,
+ "grad_norm": 0.03785555809736252,
+ "learning_rate": 0.000533195950905842,
+ "loss": 2.295558452606201,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.030696166679263115,
+ "learning_rate": 0.0005330274088790239,
+ "loss": 1.724071979522705,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 0.6137110590934753,
+ "eval_runtime": 40.4223,
+ "eval_samples_per_second": 60.412,
+ "eval_steps_per_second": 0.495,
+ "step": 864
+ },
+ {
+ "epoch": 12.013937282229966,
+ "grad_norm": 0.0449201837182045,
+ "learning_rate": 0.0005328586815232773,
+ "loss": 2.2556076049804688,
+ "step": 865
+ },
+ {
+ "epoch": 12.02787456445993,
+ "grad_norm": 0.05987463518977165,
+ "learning_rate": 0.0005326897689745428,
+ "loss": 2.2625327110290527,
+ "step": 866
+ },
+ {
+ "epoch": 12.041811846689896,
+ "grad_norm": 0.06580447405576706,
+ "learning_rate": 0.0005325206713689101,
+ "loss": 2.2511653900146484,
+ "step": 867
+ },
+ {
+ "epoch": 12.05574912891986,
+ "grad_norm": 0.08316071331501007,
+ "learning_rate": 0.0005323513888426181,
+ "loss": 2.2627158164978027,
+ "step": 868
+ },
+ {
+ "epoch": 12.069686411149826,
+ "grad_norm": 0.0916237011551857,
+ "learning_rate": 0.0005321819215320546,
+ "loss": 2.269570827484131,
+ "step": 869
+ },
+ {
+ "epoch": 12.08362369337979,
+ "grad_norm": 0.08639268577098846,
+ "learning_rate": 0.0005320122695737565,
+ "loss": 2.2877602577209473,
+ "step": 870
+ },
+ {
+ "epoch": 12.097560975609756,
+ "grad_norm": 0.08962945640087128,
+ "learning_rate": 0.0005318424331044094,
+ "loss": 2.274688959121704,
+ "step": 871
+ },
+ {
+ "epoch": 12.111498257839722,
+ "grad_norm": 0.09136641025543213,
+ "learning_rate": 0.0005316724122608473,
+ "loss": 2.277164936065674,
+ "step": 872
+ },
+ {
+ "epoch": 12.125435540069686,
+ "grad_norm": 0.08773082494735718,
+ "learning_rate": 0.0005315022071800527,
+ "loss": 2.260807991027832,
+ "step": 873
+ },
+ {
+ "epoch": 12.139372822299652,
+ "grad_norm": 0.09494079649448395,
+ "learning_rate": 0.0005313318179991574,
+ "loss": 2.268434524536133,
+ "step": 874
+ },
+ {
+ "epoch": 12.153310104529616,
+ "grad_norm": 0.08850299566984177,
+ "learning_rate": 0.0005311612448554402,
+ "loss": 2.2728161811828613,
+ "step": 875
+ },
+ {
+ "epoch": 12.167247386759582,
+ "grad_norm": 0.07735221832990646,
+ "learning_rate": 0.0005309904878863293,
+ "loss": 2.2735438346862793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181184668989546,
+ "grad_norm": 0.07927414774894714,
+ "learning_rate": 0.0005308195472294001,
+ "loss": 2.2843785285949707,
+ "step": 877
+ },
+ {
+ "epoch": 12.195121951219512,
+ "grad_norm": 0.08350995928049088,
+ "learning_rate": 0.0005306484230223769,
+ "loss": 2.2791378498077393,
+ "step": 878
+ },
+ {
+ "epoch": 12.209059233449477,
+ "grad_norm": 0.0768347680568695,
+ "learning_rate": 0.000530477115403131,
+ "loss": 2.2607197761535645,
+ "step": 879
+ },
+ {
+ "epoch": 12.222996515679442,
+ "grad_norm": 0.07743900269269943,
+ "learning_rate": 0.0005303056245096817,
+ "loss": 2.291794776916504,
+ "step": 880
+ },
+ {
+ "epoch": 12.236933797909408,
+ "grad_norm": 0.07712893933057785,
+ "learning_rate": 0.0005301339504801967,
+ "loss": 2.279090404510498,
+ "step": 881
+ },
+ {
+ "epoch": 12.250871080139373,
+ "grad_norm": 0.06713975220918655,
+ "learning_rate": 0.0005299620934529902,
+ "loss": 2.2770416736602783,
+ "step": 882
+ },
+ {
+ "epoch": 12.264808362369338,
+ "grad_norm": 0.06651429831981659,
+ "learning_rate": 0.0005297900535665244,
+ "loss": 2.275513172149658,
+ "step": 883
+ },
+ {
+ "epoch": 12.278745644599303,
+ "grad_norm": 0.07272221148014069,
+ "learning_rate": 0.0005296178309594087,
+ "loss": 2.257089614868164,
+ "step": 884
+ },
+ {
+ "epoch": 12.292682926829269,
+ "grad_norm": 0.07249968498945236,
+ "learning_rate": 0.0005294454257704,
+ "loss": 2.285494089126587,
+ "step": 885
+ },
+ {
+ "epoch": 12.306620209059233,
+ "grad_norm": 0.07468163967132568,
+ "learning_rate": 0.0005292728381384018,
+ "loss": 2.277432918548584,
+ "step": 886
+ },
+ {
+ "epoch": 12.320557491289199,
+ "grad_norm": 0.06479580700397491,
+ "learning_rate": 0.0005291000682024647,
+ "loss": 2.2599823474884033,
+ "step": 887
+ },
+ {
+ "epoch": 12.334494773519165,
+ "grad_norm": 0.064862459897995,
+ "learning_rate": 0.0005289271161017867,
+ "loss": 2.276655912399292,
+ "step": 888
+ },
+ {
+ "epoch": 12.348432055749129,
+ "grad_norm": 0.06486142426729202,
+ "learning_rate": 0.0005287539819757119,
+ "loss": 2.2744534015655518,
+ "step": 889
+ },
+ {
+ "epoch": 12.362369337979095,
+ "grad_norm": 0.05824895203113556,
+ "learning_rate": 0.0005285806659637315,
+ "loss": 2.300732135772705,
+ "step": 890
+ },
+ {
+ "epoch": 12.376306620209059,
+ "grad_norm": 0.05379778891801834,
+ "learning_rate": 0.000528407168205483,
+ "loss": 2.278594493865967,
+ "step": 891
+ },
+ {
+ "epoch": 12.390243902439025,
+ "grad_norm": 0.04621826857328415,
+ "learning_rate": 0.0005282334888407503,
+ "loss": 2.277547836303711,
+ "step": 892
+ },
+ {
+ "epoch": 12.404181184668989,
+ "grad_norm": 0.049979664385318756,
+ "learning_rate": 0.000528059628009464,
+ "loss": 2.2680468559265137,
+ "step": 893
+ },
+ {
+ "epoch": 12.418118466898955,
+ "grad_norm": 0.04380398988723755,
+ "learning_rate": 0.0005278855858517003,
+ "loss": 2.267700672149658,
+ "step": 894
+ },
+ {
+ "epoch": 12.43205574912892,
+ "grad_norm": 0.04196985065937042,
+ "learning_rate": 0.0005277113625076822,
+ "loss": 2.2651240825653076,
+ "step": 895
+ },
+ {
+ "epoch": 12.445993031358885,
+ "grad_norm": 0.04277926683425903,
+ "learning_rate": 0.000527536958117778,
+ "loss": 2.266249179840088,
+ "step": 896
+ },
+ {
+ "epoch": 12.45993031358885,
+ "grad_norm": 0.0408395491540432,
+ "learning_rate": 0.0005273623728225022,
+ "loss": 2.2429051399230957,
+ "step": 897
+ },
+ {
+ "epoch": 12.473867595818815,
+ "grad_norm": 0.043364666402339935,
+ "learning_rate": 0.0005271876067625153,
+ "loss": 2.2852165699005127,
+ "step": 898
+ },
+ {
+ "epoch": 12.487804878048781,
+ "grad_norm": 0.041563525795936584,
+ "learning_rate": 0.0005270126600786228,
+ "loss": 2.2763617038726807,
+ "step": 899
+ },
+ {
+ "epoch": 12.501742160278745,
+ "grad_norm": 0.04007960110902786,
+ "learning_rate": 0.0005268375329117762,
+ "loss": 2.2460227012634277,
+ "step": 900
+ },
+ {
+ "epoch": 12.515679442508711,
+ "grad_norm": 0.041913922876119614,
+ "learning_rate": 0.0005266622254030724,
+ "loss": 2.2742671966552734,
+ "step": 901
+ },
+ {
+ "epoch": 12.529616724738675,
+ "grad_norm": 0.03877204284071922,
+ "learning_rate": 0.0005264867376937534,
+ "loss": 2.2731478214263916,
+ "step": 902
+ },
+ {
+ "epoch": 12.543554006968641,
+ "grad_norm": 0.04184960573911667,
+ "learning_rate": 0.0005263110699252065,
+ "loss": 2.266854763031006,
+ "step": 903
+ },
+ {
+ "epoch": 12.557491289198607,
+ "grad_norm": 0.040446311235427856,
+ "learning_rate": 0.000526135222238964,
+ "loss": 2.2761149406433105,
+ "step": 904
+ },
+ {
+ "epoch": 12.571428571428571,
+ "grad_norm": 0.04059732332825661,
+ "learning_rate": 0.0005259591947767032,
+ "loss": 2.255995273590088,
+ "step": 905
+ },
+ {
+ "epoch": 12.585365853658537,
+ "grad_norm": 0.03913392499089241,
+ "learning_rate": 0.0005257829876802464,
+ "loss": 2.2813472747802734,
+ "step": 906
+ },
+ {
+ "epoch": 12.599303135888501,
+ "grad_norm": 0.038668543100357056,
+ "learning_rate": 0.0005256066010915603,
+ "loss": 2.267693519592285,
+ "step": 907
+ },
+ {
+ "epoch": 12.613240418118467,
+ "grad_norm": 0.03873024880886078,
+ "learning_rate": 0.0005254300351527566,
+ "loss": 2.2710647583007812,
+ "step": 908
+ },
+ {
+ "epoch": 12.627177700348431,
+ "grad_norm": 0.03990354388952255,
+ "learning_rate": 0.000525253290006091,
+ "loss": 2.272221088409424,
+ "step": 909
+ },
+ {
+ "epoch": 12.641114982578397,
+ "grad_norm": 0.037710171192884445,
+ "learning_rate": 0.0005250763657939641,
+ "loss": 2.285928249359131,
+ "step": 910
+ },
+ {
+ "epoch": 12.655052264808361,
+ "grad_norm": 0.03865284472703934,
+ "learning_rate": 0.0005248992626589203,
+ "loss": 2.2687816619873047,
+ "step": 911
+ },
+ {
+ "epoch": 12.668989547038327,
+ "grad_norm": 0.040922585874795914,
+ "learning_rate": 0.0005247219807436489,
+ "loss": 2.278773546218872,
+ "step": 912
+ },
+ {
+ "epoch": 12.682926829268293,
+ "grad_norm": 0.0420064739882946,
+ "learning_rate": 0.000524544520190982,
+ "loss": 2.28548002243042,
+ "step": 913
+ },
+ {
+ "epoch": 12.696864111498257,
+ "grad_norm": 0.039056528359651566,
+ "learning_rate": 0.000524366881143897,
+ "loss": 2.2802648544311523,
+ "step": 914
+ },
+ {
+ "epoch": 12.710801393728223,
+ "grad_norm": 0.037798572331666946,
+ "learning_rate": 0.0005241890637455141,
+ "loss": 2.272362470626831,
+ "step": 915
+ },
+ {
+ "epoch": 12.724738675958188,
+ "grad_norm": 0.03791450709104538,
+ "learning_rate": 0.0005240110681390978,
+ "loss": 2.289794921875,
+ "step": 916
+ },
+ {
+ "epoch": 12.738675958188153,
+ "grad_norm": 0.03728414699435234,
+ "learning_rate": 0.0005238328944680558,
+ "loss": 2.2741048336029053,
+ "step": 917
+ },
+ {
+ "epoch": 12.752613240418118,
+ "grad_norm": 0.03752773255109787,
+ "learning_rate": 0.0005236545428759394,
+ "loss": 2.265214443206787,
+ "step": 918
+ },
+ {
+ "epoch": 12.766550522648084,
+ "grad_norm": 0.0390285924077034,
+ "learning_rate": 0.0005234760135064434,
+ "loss": 2.2683982849121094,
+ "step": 919
+ },
+ {
+ "epoch": 12.78048780487805,
+ "grad_norm": 0.037184618413448334,
+ "learning_rate": 0.0005232973065034055,
+ "loss": 2.2899575233459473,
+ "step": 920
+ },
+ {
+ "epoch": 12.794425087108014,
+ "grad_norm": 0.036621104925870895,
+ "learning_rate": 0.0005231184220108069,
+ "loss": 2.2683801651000977,
+ "step": 921
+ },
+ {
+ "epoch": 12.80836236933798,
+ "grad_norm": 0.03748784959316254,
+ "learning_rate": 0.0005229393601727713,
+ "loss": 2.292527914047241,
+ "step": 922
+ },
+ {
+ "epoch": 12.822299651567944,
+ "grad_norm": 0.038282785564661026,
+ "learning_rate": 0.000522760121133566,
+ "loss": 2.2873587608337402,
+ "step": 923
+ },
+ {
+ "epoch": 12.83623693379791,
+ "grad_norm": 0.03690093755722046,
+ "learning_rate": 0.0005225807050376004,
+ "loss": 2.278165340423584,
+ "step": 924
+ },
+ {
+ "epoch": 12.850174216027874,
+ "grad_norm": 0.036749616265296936,
+ "learning_rate": 0.0005224011120294269,
+ "loss": 2.2742109298706055,
+ "step": 925
+ },
+ {
+ "epoch": 12.86411149825784,
+ "grad_norm": 0.03896362707018852,
+ "learning_rate": 0.0005222213422537401,
+ "loss": 2.268939971923828,
+ "step": 926
+ },
+ {
+ "epoch": 12.878048780487806,
+ "grad_norm": 0.03734518960118294,
+ "learning_rate": 0.0005220413958553777,
+ "loss": 2.2679965496063232,
+ "step": 927
+ },
+ {
+ "epoch": 12.89198606271777,
+ "grad_norm": 0.03745289146900177,
+ "learning_rate": 0.0005218612729793188,
+ "loss": 2.281343460083008,
+ "step": 928
+ },
+ {
+ "epoch": 12.905923344947736,
+ "grad_norm": 0.037503164261579514,
+ "learning_rate": 0.0005216809737706856,
+ "loss": 2.253847599029541,
+ "step": 929
+ },
+ {
+ "epoch": 12.9198606271777,
+ "grad_norm": 0.03740404173731804,
+ "learning_rate": 0.0005215004983747415,
+ "loss": 2.286041259765625,
+ "step": 930
+ },
+ {
+ "epoch": 12.933797909407666,
+ "grad_norm": 0.038288187235593796,
+ "learning_rate": 0.0005213198469368925,
+ "loss": 2.2621493339538574,
+ "step": 931
+ },
+ {
+ "epoch": 12.94773519163763,
+ "grad_norm": 0.037316542118787766,
+ "learning_rate": 0.0005211390196026861,
+ "loss": 2.2640275955200195,
+ "step": 932
+ },
+ {
+ "epoch": 12.961672473867596,
+ "grad_norm": 0.036008067429065704,
+ "learning_rate": 0.0005209580165178117,
+ "loss": 2.2758469581604004,
+ "step": 933
+ },
+ {
+ "epoch": 12.975609756097562,
+ "grad_norm": 0.035242367535829544,
+ "learning_rate": 0.0005207768378281003,
+ "loss": 2.272761344909668,
+ "step": 934
+ },
+ {
+ "epoch": 12.989547038327526,
+ "grad_norm": 0.03540848195552826,
+ "learning_rate": 0.0005205954836795241,
+ "loss": 2.265636444091797,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.02824031375348568,
+ "learning_rate": 0.000520413954218197,
+ "loss": 1.7006406784057617,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 0.6141775250434875,
+ "eval_runtime": 40.6243,
+ "eval_samples_per_second": 60.112,
+ "eval_steps_per_second": 0.492,
+ "step": 936
+ },
+ {
+ "epoch": 13.013937282229966,
+ "grad_norm": 0.04125862568616867,
+ "learning_rate": 0.0005202322495903739,
+ "loss": 2.215397357940674,
+ "step": 937
+ },
+ {
+ "epoch": 13.02787456445993,
+ "grad_norm": 0.05152325704693794,
+ "learning_rate": 0.0005200503699424509,
+ "loss": 2.2176401615142822,
+ "step": 938
+ },
+ {
+ "epoch": 13.041811846689896,
+ "grad_norm": 0.053824957460165024,
+ "learning_rate": 0.0005198683154209654,
+ "loss": 2.224720001220703,
+ "step": 939
+ },
+ {
+ "epoch": 13.05574912891986,
+ "grad_norm": 0.06314096599817276,
+ "learning_rate": 0.0005196860861725952,
+ "loss": 2.2176055908203125,
+ "step": 940
+ },
+ {
+ "epoch": 13.069686411149826,
+ "grad_norm": 0.06466981023550034,
+ "learning_rate": 0.0005195036823441592,
+ "loss": 2.2175378799438477,
+ "step": 941
+ },
+ {
+ "epoch": 13.08362369337979,
+ "grad_norm": 0.06475649029016495,
+ "learning_rate": 0.0005193211040826169,
+ "loss": 2.227267265319824,
+ "step": 942
+ },
+ {
+ "epoch": 13.097560975609756,
+ "grad_norm": 0.07283888757228851,
+ "learning_rate": 0.0005191383515350681,
+ "loss": 2.216702938079834,
+ "step": 943
+ },
+ {
+ "epoch": 13.111498257839722,
+ "grad_norm": 0.0799742266535759,
+ "learning_rate": 0.0005189554248487536,
+ "loss": 2.2331976890563965,
+ "step": 944
+ },
+ {
+ "epoch": 13.125435540069686,
+ "grad_norm": 0.07081552594900131,
+ "learning_rate": 0.0005187723241710539,
+ "loss": 2.210658550262451,
+ "step": 945
+ },
+ {
+ "epoch": 13.139372822299652,
+ "grad_norm": 0.07200472056865692,
+ "learning_rate": 0.0005185890496494899,
+ "loss": 2.239154815673828,
+ "step": 946
+ },
+ {
+ "epoch": 13.153310104529616,
+ "grad_norm": 0.06486581265926361,
+ "learning_rate": 0.0005184056014317225,
+ "loss": 2.2305479049682617,
+ "step": 947
+ },
+ {
+ "epoch": 13.167247386759582,
+ "grad_norm": 0.06277933716773987,
+ "learning_rate": 0.0005182219796655528,
+ "loss": 2.244959831237793,
+ "step": 948
+ },
+ {
+ "epoch": 13.181184668989546,
+ "grad_norm": 0.06154560670256615,
+ "learning_rate": 0.0005180381844989214,
+ "loss": 2.2427070140838623,
+ "step": 949
+ },
+ {
+ "epoch": 13.195121951219512,
+ "grad_norm": 0.05721663311123848,
+ "learning_rate": 0.0005178542160799088,
+ "loss": 2.2324533462524414,
+ "step": 950
+ },
+ {
+ "epoch": 13.209059233449477,
+ "grad_norm": 0.06185296177864075,
+ "learning_rate": 0.0005176700745567353,
+ "loss": 2.2416887283325195,
+ "step": 951
+ },
+ {
+ "epoch": 13.222996515679442,
+ "grad_norm": 0.05919089540839195,
+ "learning_rate": 0.0005174857600777599,
+ "loss": 2.2196006774902344,
+ "step": 952
+ },
+ {
+ "epoch": 13.236933797909408,
+ "grad_norm": 0.0566907562315464,
+ "learning_rate": 0.0005173012727914816,
+ "loss": 2.2388274669647217,
+ "step": 953
+ },
+ {
+ "epoch": 13.250871080139373,
+ "grad_norm": 0.05523713305592537,
+ "learning_rate": 0.0005171166128465389,
+ "loss": 2.229844093322754,
+ "step": 954
+ },
+ {
+ "epoch": 13.264808362369338,
+ "grad_norm": 0.05697935074567795,
+ "learning_rate": 0.0005169317803917083,
+ "loss": 2.236034393310547,
+ "step": 955
+ },
+ {
+ "epoch": 13.278745644599303,
+ "grad_norm": 0.05645161122083664,
+ "learning_rate": 0.0005167467755759065,
+ "loss": 2.2360434532165527,
+ "step": 956
+ },
+ {
+ "epoch": 13.292682926829269,
+ "grad_norm": 0.0592060424387455,
+ "learning_rate": 0.0005165615985481881,
+ "loss": 2.24061918258667,
+ "step": 957
+ },
+ {
+ "epoch": 13.306620209059233,
+ "grad_norm": 0.060027722269296646,
+ "learning_rate": 0.0005163762494577474,
+ "loss": 2.2592904567718506,
+ "step": 958
+ },
+ {
+ "epoch": 13.320557491289199,
+ "grad_norm": 0.05506884679198265,
+ "learning_rate": 0.0005161907284539165,
+ "loss": 2.2442619800567627,
+ "step": 959
+ },
+ {
+ "epoch": 13.334494773519165,
+ "grad_norm": 0.05131617188453674,
+ "learning_rate": 0.0005160050356861663,
+ "loss": 2.2405781745910645,
+ "step": 960
+ },
+ {
+ "epoch": 13.348432055749129,
+ "grad_norm": 0.04664314538240433,
+ "learning_rate": 0.0005158191713041063,
+ "loss": 2.2319693565368652,
+ "step": 961
+ },
+ {
+ "epoch": 13.362369337979095,
+ "grad_norm": 0.04470345377922058,
+ "learning_rate": 0.0005156331354574838,
+ "loss": 2.237271785736084,
+ "step": 962
+ },
+ {
+ "epoch": 13.376306620209059,
+ "grad_norm": 0.04382263496518135,
+ "learning_rate": 0.0005154469282961847,
+ "loss": 2.2390639781951904,
+ "step": 963
+ },
+ {
+ "epoch": 13.390243902439025,
+ "grad_norm": 0.046823810786008835,
+ "learning_rate": 0.0005152605499702328,
+ "loss": 2.249532699584961,
+ "step": 964
+ },
+ {
+ "epoch": 13.404181184668989,
+ "grad_norm": 0.047521062195301056,
+ "learning_rate": 0.0005150740006297896,
+ "loss": 2.2330965995788574,
+ "step": 965
+ },
+ {
+ "epoch": 13.418118466898955,
+ "grad_norm": 0.04479820281267166,
+ "learning_rate": 0.0005148872804251548,
+ "loss": 2.251351833343506,
+ "step": 966
+ },
+ {
+ "epoch": 13.43205574912892,
+ "grad_norm": 0.04384589567780495,
+ "learning_rate": 0.000514700389506765,
+ "loss": 2.2227377891540527,
+ "step": 967
+ },
+ {
+ "epoch": 13.445993031358885,
+ "grad_norm": 0.04421713575720787,
+ "learning_rate": 0.0005145133280251954,
+ "loss": 2.240830421447754,
+ "step": 968
+ },
+ {
+ "epoch": 13.45993031358885,
+ "grad_norm": 0.04206979274749756,
+ "learning_rate": 0.0005143260961311578,
+ "loss": 2.2531075477600098,
+ "step": 969
+ },
+ {
+ "epoch": 13.473867595818815,
+ "grad_norm": 0.04159024357795715,
+ "learning_rate": 0.0005141386939755014,
+ "loss": 2.2377514839172363,
+ "step": 970
+ },
+ {
+ "epoch": 13.487804878048781,
+ "grad_norm": 0.04292576014995575,
+ "learning_rate": 0.0005139511217092129,
+ "loss": 2.227909564971924,
+ "step": 971
+ },
+ {
+ "epoch": 13.501742160278745,
+ "grad_norm": 0.044184450060129166,
+ "learning_rate": 0.000513763379483416,
+ "loss": 2.2394638061523438,
+ "step": 972
+ },
+ {
+ "epoch": 13.515679442508711,
+ "grad_norm": 0.044156551361083984,
+ "learning_rate": 0.0005135754674493709,
+ "loss": 2.2507641315460205,
+ "step": 973
+ },
+ {
+ "epoch": 13.529616724738675,
+ "grad_norm": 0.04301406443119049,
+ "learning_rate": 0.0005133873857584753,
+ "loss": 2.228238582611084,
+ "step": 974
+ },
+ {
+ "epoch": 13.543554006968641,
+ "grad_norm": 0.044542353600263596,
+ "learning_rate": 0.0005131991345622629,
+ "loss": 2.233346700668335,
+ "step": 975
+ },
+ {
+ "epoch": 13.557491289198607,
+ "grad_norm": 0.04516741260886192,
+ "learning_rate": 0.0005130107140124046,
+ "loss": 2.257772922515869,
+ "step": 976
+ },
+ {
+ "epoch": 13.571428571428571,
+ "grad_norm": 0.04428017884492874,
+ "learning_rate": 0.0005128221242607072,
+ "loss": 2.2487168312072754,
+ "step": 977
+ },
+ {
+ "epoch": 13.585365853658537,
+ "grad_norm": 0.04395915940403938,
+ "learning_rate": 0.0005126333654591141,
+ "loss": 2.2402656078338623,
+ "step": 978
+ },
+ {
+ "epoch": 13.599303135888501,
+ "grad_norm": 0.04147777706384659,
+ "learning_rate": 0.0005124444377597049,
+ "loss": 2.249999523162842,
+ "step": 979
+ },
+ {
+ "epoch": 13.613240418118467,
+ "grad_norm": 0.04329652711749077,
+ "learning_rate": 0.0005122553413146951,
+ "loss": 2.23622465133667,
+ "step": 980
+ },
+ {
+ "epoch": 13.627177700348431,
+ "grad_norm": 0.04756402596831322,
+ "learning_rate": 0.0005120660762764366,
+ "loss": 2.236121892929077,
+ "step": 981
+ },
+ {
+ "epoch": 13.641114982578397,
+ "grad_norm": 0.0474284365773201,
+ "learning_rate": 0.0005118766427974166,
+ "loss": 2.255303144454956,
+ "step": 982
+ },
+ {
+ "epoch": 13.655052264808361,
+ "grad_norm": 0.043167296797037125,
+ "learning_rate": 0.0005116870410302582,
+ "loss": 2.2689664363861084,
+ "step": 983
+ },
+ {
+ "epoch": 13.668989547038327,
+ "grad_norm": 0.0410541407763958,
+ "learning_rate": 0.0005114972711277202,
+ "loss": 2.2468833923339844,
+ "step": 984
+ },
+ {
+ "epoch": 13.682926829268293,
+ "grad_norm": 0.04530128464102745,
+ "learning_rate": 0.0005113073332426968,
+ "loss": 2.252777576446533,
+ "step": 985
+ },
+ {
+ "epoch": 13.696864111498257,
+ "grad_norm": 0.047060899436473846,
+ "learning_rate": 0.0005111172275282174,
+ "loss": 2.2606921195983887,
+ "step": 986
+ },
+ {
+ "epoch": 13.710801393728223,
+ "grad_norm": 0.046014782041311264,
+ "learning_rate": 0.0005109269541374469,
+ "loss": 2.2331900596618652,
+ "step": 987
+ },
+ {
+ "epoch": 13.724738675958188,
+ "grad_norm": 0.04229343682527542,
+ "learning_rate": 0.000510736513223685,
+ "loss": 2.2451624870300293,
+ "step": 988
+ },
+ {
+ "epoch": 13.738675958188153,
+ "grad_norm": 0.04296563193202019,
+ "learning_rate": 0.0005105459049403665,
+ "loss": 2.25291109085083,
+ "step": 989
+ },
+ {
+ "epoch": 13.752613240418118,
+ "grad_norm": 0.04043250158429146,
+ "learning_rate": 0.0005103551294410614,
+ "loss": 2.251645088195801,
+ "step": 990
+ },
+ {
+ "epoch": 13.766550522648084,
+ "grad_norm": 0.04019814357161522,
+ "learning_rate": 0.0005101641868794736,
+ "loss": 2.2456679344177246,
+ "step": 991
+ },
+ {
+ "epoch": 13.78048780487805,
+ "grad_norm": 0.039857521653175354,
+ "learning_rate": 0.0005099730774094423,
+ "loss": 2.2425897121429443,
+ "step": 992
+ },
+ {
+ "epoch": 13.794425087108014,
+ "grad_norm": 0.039773937314748764,
+ "learning_rate": 0.0005097818011849408,
+ "loss": 2.2354488372802734,
+ "step": 993
+ },
+ {
+ "epoch": 13.80836236933798,
+ "grad_norm": 0.04233181104063988,
+ "learning_rate": 0.0005095903583600772,
+ "loss": 2.234449863433838,
+ "step": 994
+ },
+ {
+ "epoch": 13.822299651567944,
+ "grad_norm": 0.03733415529131889,
+ "learning_rate": 0.0005093987490890933,
+ "loss": 2.2579874992370605,
+ "step": 995
+ },
+ {
+ "epoch": 13.83623693379791,
+ "grad_norm": 0.039548005908727646,
+ "learning_rate": 0.0005092069735263651,
+ "loss": 2.2390646934509277,
+ "step": 996
+ },
+ {
+ "epoch": 13.850174216027874,
+ "grad_norm": 0.03964848443865776,
+ "learning_rate": 0.000509015031826403,
+ "loss": 2.245936393737793,
+ "step": 997
+ },
+ {
+ "epoch": 13.86411149825784,
+ "grad_norm": 0.0411066859960556,
+ "learning_rate": 0.0005088229241438509,
+ "loss": 2.248837471008301,
+ "step": 998
+ },
+ {
+ "epoch": 13.878048780487806,
+ "grad_norm": 0.042689017951488495,
+ "learning_rate": 0.0005086306506334864,
+ "loss": 2.2478384971618652,
+ "step": 999
+ },
+ {
+ "epoch": 13.89198606271777,
+ "grad_norm": 0.04124174639582634,
+ "learning_rate": 0.0005084382114502208,
+ "loss": 2.242241382598877,
+ "step": 1000
+ },
+ {
+ "epoch": 13.905923344947736,
+ "grad_norm": 0.03894231840968132,
+ "learning_rate": 0.0005082456067490988,
+ "loss": 2.2491180896759033,
+ "step": 1001
+ },
+ {
+ "epoch": 13.9198606271777,
+ "grad_norm": 0.04093318060040474,
+ "learning_rate": 0.0005080528366852987,
+ "loss": 2.2608485221862793,
+ "step": 1002
+ },
+ {
+ "epoch": 13.933797909407666,
+ "grad_norm": 0.040943752974271774,
+ "learning_rate": 0.0005078599014141318,
+ "loss": 2.246333360671997,
+ "step": 1003
+ },
+ {
+ "epoch": 13.94773519163763,
+ "grad_norm": 0.04095872491598129,
+ "learning_rate": 0.0005076668010910425,
+ "loss": 2.2806015014648438,
+ "step": 1004
+ },
+ {
+ "epoch": 13.961672473867596,
+ "grad_norm": 0.042693428695201874,
+ "learning_rate": 0.0005074735358716083,
+ "loss": 2.2343358993530273,
+ "step": 1005
+ },
+ {
+ "epoch": 13.975609756097562,
+ "grad_norm": 0.04350119084119797,
+ "learning_rate": 0.0005072801059115394,
+ "loss": 2.258847713470459,
+ "step": 1006
+ },
+ {
+ "epoch": 13.989547038327526,
+ "grad_norm": 0.03934241831302643,
+ "learning_rate": 0.000507086511366679,
+ "loss": 2.2553815841674805,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.031480252742767334,
+ "learning_rate": 0.0005068927523930027,
+ "loss": 1.6732367277145386,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 0.614798903465271,
+ "eval_runtime": 41.0771,
+ "eval_samples_per_second": 59.449,
+ "eval_steps_per_second": 0.487,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013937282229966,
+ "grad_norm": 0.04352101311087608,
+ "learning_rate": 0.0005066988291466184,
+ "loss": 2.189373254776001,
+ "step": 1009
+ },
+ {
+ "epoch": 14.02787456445993,
+ "grad_norm": 0.058463841676712036,
+ "learning_rate": 0.0005065047417837669,
+ "loss": 2.2095961570739746,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041811846689896,
+ "grad_norm": 0.06910230219364166,
+ "learning_rate": 0.0005063104904608206,
+ "loss": 2.1949195861816406,
+ "step": 1011
+ },
+ {
+ "epoch": 14.05574912891986,
+ "grad_norm": 0.07689601182937622,
+ "learning_rate": 0.0005061160753342843,
+ "loss": 2.21500563621521,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069686411149826,
+ "grad_norm": 0.0781058669090271,
+ "learning_rate": 0.0005059214965607948,
+ "loss": 2.2030768394470215,
+ "step": 1013
+ },
+ {
+ "epoch": 14.08362369337979,
+ "grad_norm": 0.08622978627681732,
+ "learning_rate": 0.0005057267542971207,
+ "loss": 2.209932327270508,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097560975609756,
+ "grad_norm": 0.08425556868314743,
+ "learning_rate": 0.0005055318487001624,
+ "loss": 2.2136945724487305,
+ "step": 1015
+ },
+ {
+ "epoch": 14.111498257839722,
+ "grad_norm": 0.07886914908885956,
+ "learning_rate": 0.0005053367799269518,
+ "loss": 2.2054924964904785,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125435540069686,
+ "grad_norm": 0.07874055951833725,
+ "learning_rate": 0.0005051415481346522,
+ "loss": 2.183264970779419,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139372822299652,
+ "grad_norm": 0.07370466738939285,
+ "learning_rate": 0.0005049461534805585,
+ "loss": 2.207981586456299,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153310104529616,
+ "grad_norm": 0.07387633621692657,
+ "learning_rate": 0.0005047505961220964,
+ "loss": 2.2039785385131836,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167247386759582,
+ "grad_norm": 0.08056999742984772,
+ "learning_rate": 0.0005045548762168231,
+ "loss": 2.1992383003234863,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181184668989546,
+ "grad_norm": 0.08819372206926346,
+ "learning_rate": 0.0005043589939224266,
+ "loss": 2.2112417221069336,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195121951219512,
+ "grad_norm": 0.09344980865716934,
+ "learning_rate": 0.0005041629493967257,
+ "loss": 2.2054364681243896,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209059233449477,
+ "grad_norm": 0.08366914838552475,
+ "learning_rate": 0.0005039667427976697,
+ "loss": 2.2166051864624023,
+ "step": 1023
+ },
+ {
+ "epoch": 14.222996515679442,
+ "grad_norm": 0.08368150889873505,
+ "learning_rate": 0.0005037703742833391,
+ "loss": 2.2117366790771484,
+ "step": 1024
+ },
+ {
+ "epoch": 14.236933797909408,
+ "grad_norm": 0.07882174849510193,
+ "learning_rate": 0.0005035738440119442,
+ "loss": 2.210777759552002,
+ "step": 1025
+ },
+ {
+ "epoch": 14.250871080139373,
+ "grad_norm": 0.06474770605564117,
+ "learning_rate": 0.0005033771521418261,
+ "loss": 2.2155799865722656,
+ "step": 1026
+ },
+ {
+ "epoch": 14.264808362369338,
+ "grad_norm": 0.06636861711740494,
+ "learning_rate": 0.0005031802988314557,
+ "loss": 2.212151527404785,
+ "step": 1027
+ },
+ {
+ "epoch": 14.278745644599303,
+ "grad_norm": 0.06055188179016113,
+ "learning_rate": 0.0005029832842394342,
+ "loss": 2.1962785720825195,
+ "step": 1028
+ },
+ {
+ "epoch": 14.292682926829269,
+ "grad_norm": 0.06111276522278786,
+ "learning_rate": 0.0005027861085244927,
+ "loss": 2.196725606918335,
+ "step": 1029
+ },
+ {
+ "epoch": 14.306620209059233,
+ "grad_norm": 0.06042694300413132,
+ "learning_rate": 0.0005025887718454922,
+ "loss": 2.200380325317383,
+ "step": 1030
+ },
+ {
+ "epoch": 14.320557491289199,
+ "grad_norm": 0.06327551603317261,
+ "learning_rate": 0.0005023912743614232,
+ "loss": 2.2274162769317627,
+ "step": 1031
+ },
+ {
+ "epoch": 14.334494773519165,
+ "grad_norm": 0.05598565563559532,
+ "learning_rate": 0.0005021936162314058,
+ "loss": 2.1938066482543945,
+ "step": 1032
+ },
+ {
+ "epoch": 14.348432055749129,
+ "grad_norm": 0.052598968148231506,
+ "learning_rate": 0.0005019957976146898,
+ "loss": 2.2106595039367676,
+ "step": 1033
+ },
+ {
+ "epoch": 14.362369337979095,
+ "grad_norm": 0.051847606897354126,
+ "learning_rate": 0.0005017978186706538,
+ "loss": 2.2053873538970947,
+ "step": 1034
+ },
+ {
+ "epoch": 14.376306620209059,
+ "grad_norm": 0.049158211797475815,
+ "learning_rate": 0.0005015996795588062,
+ "loss": 2.2248616218566895,
+ "step": 1035
+ },
+ {
+ "epoch": 14.390243902439025,
+ "grad_norm": 0.04686322063207626,
+ "learning_rate": 0.0005014013804387838,
+ "loss": 2.2308197021484375,
+ "step": 1036
+ },
+ {
+ "epoch": 14.404181184668989,
+ "grad_norm": 0.04794236645102501,
+ "learning_rate": 0.0005012029214703529,
+ "loss": 2.2123966217041016,
+ "step": 1037
+ },
+ {
+ "epoch": 14.418118466898955,
+ "grad_norm": 0.0469527393579483,
+ "learning_rate": 0.000501004302813408,
+ "loss": 2.2099480628967285,
+ "step": 1038
+ },
+ {
+ "epoch": 14.43205574912892,
+ "grad_norm": 0.04573623463511467,
+ "learning_rate": 0.0005008055246279727,
+ "loss": 2.2187681198120117,
+ "step": 1039
+ },
+ {
+ "epoch": 14.445993031358885,
+ "grad_norm": 0.04652945324778557,
+ "learning_rate": 0.000500606587074199,
+ "loss": 2.2252659797668457,
+ "step": 1040
+ },
+ {
+ "epoch": 14.45993031358885,
+ "grad_norm": 0.049546610563993454,
+ "learning_rate": 0.0005004074903123674,
+ "loss": 2.227400779724121,
+ "step": 1041
+ },
+ {
+ "epoch": 14.473867595818815,
+ "grad_norm": 0.04916404187679291,
+ "learning_rate": 0.0005002082345028864,
+ "loss": 2.2279880046844482,
+ "step": 1042
+ },
+ {
+ "epoch": 14.487804878048781,
+ "grad_norm": 0.04645737633109093,
+ "learning_rate": 0.0005000088198062929,
+ "loss": 2.2298312187194824,
+ "step": 1043
+ },
+ {
+ "epoch": 14.501742160278745,
+ "grad_norm": 0.04530136659741402,
+ "learning_rate": 0.0004998092463832516,
+ "loss": 2.1884169578552246,
+ "step": 1044
+ },
+ {
+ "epoch": 14.515679442508711,
+ "grad_norm": 0.043793950229883194,
+ "learning_rate": 0.0004996095143945554,
+ "loss": 2.212510585784912,
+ "step": 1045
+ },
+ {
+ "epoch": 14.529616724738675,
+ "grad_norm": 0.044199440628290176,
+ "learning_rate": 0.0004994096240011246,
+ "loss": 2.210566759109497,
+ "step": 1046
+ },
+ {
+ "epoch": 14.543554006968641,
+ "grad_norm": 0.04696475714445114,
+ "learning_rate": 0.0004992095753640072,
+ "loss": 2.223098039627075,
+ "step": 1047
+ },
+ {
+ "epoch": 14.557491289198607,
+ "grad_norm": 0.0455484576523304,
+ "learning_rate": 0.0004990093686443791,
+ "loss": 2.2351226806640625,
+ "step": 1048
+ },
+ {
+ "epoch": 14.571428571428571,
+ "grad_norm": 0.04543624818325043,
+ "learning_rate": 0.000498809004003543,
+ "loss": 2.2208995819091797,
+ "step": 1049
+ },
+ {
+ "epoch": 14.585365853658537,
+ "grad_norm": 0.048978544771671295,
+ "learning_rate": 0.0004986084816029291,
+ "loss": 2.2327723503112793,
+ "step": 1050
+ },
+ {
+ "epoch": 14.599303135888501,
+ "grad_norm": 0.04855142533779144,
+ "learning_rate": 0.0004984078016040946,
+ "loss": 2.230057716369629,
+ "step": 1051
+ },
+ {
+ "epoch": 14.613240418118467,
+ "grad_norm": 0.048660870641469955,
+ "learning_rate": 0.000498206964168724,
+ "loss": 2.2205374240875244,
+ "step": 1052
+ },
+ {
+ "epoch": 14.627177700348431,
+ "grad_norm": 0.04402444139122963,
+ "learning_rate": 0.000498005969458628,
+ "loss": 2.2225751876831055,
+ "step": 1053
+ },
+ {
+ "epoch": 14.641114982578397,
+ "grad_norm": 0.043294575065374374,
+ "learning_rate": 0.0004978048176357447,
+ "loss": 2.2238705158233643,
+ "step": 1054
+ },
+ {
+ "epoch": 14.655052264808361,
+ "grad_norm": 0.04567226022481918,
+ "learning_rate": 0.0004976035088621385,
+ "loss": 2.2082672119140625,
+ "step": 1055
+ },
+ {
+ "epoch": 14.668989547038327,
+ "grad_norm": 0.04494187608361244,
+ "learning_rate": 0.0004974020433000003,
+ "loss": 2.2213430404663086,
+ "step": 1056
+ },
+ {
+ "epoch": 14.682926829268293,
+ "grad_norm": 0.04698248207569122,
+ "learning_rate": 0.000497200421111647,
+ "loss": 2.244389772415161,
+ "step": 1057
+ },
+ {
+ "epoch": 14.696864111498257,
+ "grad_norm": 0.044764794409275055,
+ "learning_rate": 0.0004969986424595221,
+ "loss": 2.2104992866516113,
+ "step": 1058
+ },
+ {
+ "epoch": 14.710801393728223,
+ "grad_norm": 0.04466947913169861,
+ "learning_rate": 0.0004967967075061949,
+ "loss": 2.2288198471069336,
+ "step": 1059
+ },
+ {
+ "epoch": 14.724738675958188,
+ "grad_norm": 0.043151769787073135,
+ "learning_rate": 0.0004965946164143609,
+ "loss": 2.235086441040039,
+ "step": 1060
+ },
+ {
+ "epoch": 14.738675958188153,
+ "grad_norm": 0.04314344376325607,
+ "learning_rate": 0.0004963923693468413,
+ "loss": 2.226283073425293,
+ "step": 1061
+ },
+ {
+ "epoch": 14.752613240418118,
+ "grad_norm": 0.042341820895671844,
+ "learning_rate": 0.0004961899664665827,
+ "loss": 2.2036399841308594,
+ "step": 1062
+ },
+ {
+ "epoch": 14.766550522648084,
+ "grad_norm": 0.04628158360719681,
+ "learning_rate": 0.0004959874079366576,
+ "loss": 2.2239391803741455,
+ "step": 1063
+ },
+ {
+ "epoch": 14.78048780487805,
+ "grad_norm": 0.04156630113720894,
+ "learning_rate": 0.0004957846939202637,
+ "loss": 2.2288248538970947,
+ "step": 1064
+ },
+ {
+ "epoch": 14.794425087108014,
+ "grad_norm": 0.04018588736653328,
+ "learning_rate": 0.000495581824580724,
+ "loss": 2.2364609241485596,
+ "step": 1065
+ },
+ {
+ "epoch": 14.80836236933798,
+ "grad_norm": 0.04115775600075722,
+ "learning_rate": 0.0004953788000814866,
+ "loss": 2.2389607429504395,
+ "step": 1066
+ },
+ {
+ "epoch": 14.822299651567944,
+ "grad_norm": 0.044060226529836655,
+ "learning_rate": 0.000495175620586125,
+ "loss": 2.213273525238037,
+ "step": 1067
+ },
+ {
+ "epoch": 14.83623693379791,
+ "grad_norm": 0.04164741933345795,
+ "learning_rate": 0.0004949722862583368,
+ "loss": 2.2250351905822754,
+ "step": 1068
+ },
+ {
+ "epoch": 14.850174216027874,
+ "grad_norm": 0.040345050394535065,
+ "learning_rate": 0.000494768797261945,
+ "loss": 2.2460896968841553,
+ "step": 1069
+ },
+ {
+ "epoch": 14.86411149825784,
+ "grad_norm": 0.04131777212023735,
+ "learning_rate": 0.000494565153760897,
+ "loss": 2.222750186920166,
+ "step": 1070
+ },
+ {
+ "epoch": 14.878048780487806,
+ "grad_norm": 0.043353449553251266,
+ "learning_rate": 0.0004943613559192648,
+ "loss": 2.2270917892456055,
+ "step": 1071
+ },
+ {
+ "epoch": 14.89198606271777,
+ "grad_norm": 0.040032703429460526,
+ "learning_rate": 0.0004941574039012444,
+ "loss": 2.236917495727539,
+ "step": 1072
+ },
+ {
+ "epoch": 14.905923344947736,
+ "grad_norm": 0.04349394142627716,
+ "learning_rate": 0.0004939532978711566,
+ "loss": 2.227550983428955,
+ "step": 1073
+ },
+ {
+ "epoch": 14.9198606271777,
+ "grad_norm": 0.04321689531207085,
+ "learning_rate": 0.0004937490379934456,
+ "loss": 2.224440097808838,
+ "step": 1074
+ },
+ {
+ "epoch": 14.933797909407666,
+ "grad_norm": 0.04186500608921051,
+ "learning_rate": 0.00049354462443268,
+ "loss": 2.2296669483184814,
+ "step": 1075
+ },
+ {
+ "epoch": 14.94773519163763,
+ "grad_norm": 0.04348273575305939,
+ "learning_rate": 0.0004933400573535521,
+ "loss": 2.2341816425323486,
+ "step": 1076
+ },
+ {
+ "epoch": 14.961672473867596,
+ "grad_norm": 0.042280469089746475,
+ "learning_rate": 0.000493135336920878,
+ "loss": 2.216071605682373,
+ "step": 1077
+ },
+ {
+ "epoch": 14.975609756097562,
+ "grad_norm": 0.042124394327402115,
+ "learning_rate": 0.0004929304632995973,
+ "loss": 2.234611749649048,
+ "step": 1078
+ },
+ {
+ "epoch": 14.989547038327526,
+ "grad_norm": 0.04272354766726494,
+ "learning_rate": 0.0004927254366547727,
+ "loss": 2.2185311317443848,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.03671395778656006,
+ "learning_rate": 0.0004925202571515908,
+ "loss": 1.6621125936508179,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 0.6159878969192505,
+ "eval_runtime": 40.0379,
+ "eval_samples_per_second": 60.992,
+ "eval_steps_per_second": 0.5,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013937282229966,
+ "grad_norm": 0.04342801868915558,
+ "learning_rate": 0.0004923149249553608,
+ "loss": 2.1830663681030273,
+ "step": 1081
+ },
+ {
+ "epoch": 15.02787456445993,
+ "grad_norm": 0.05104755610227585,
+ "learning_rate": 0.0004921094402315152,
+ "loss": 2.1628425121307373,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041811846689896,
+ "grad_norm": 0.056926626712083817,
+ "learning_rate": 0.0004919038031456093,
+ "loss": 2.1713647842407227,
+ "step": 1083
+ },
+ {
+ "epoch": 15.05574912891986,
+ "grad_norm": 0.06162258982658386,
+ "learning_rate": 0.0004916980138633213,
+ "loss": 2.181713342666626,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069686411149826,
+ "grad_norm": 0.06958366185426712,
+ "learning_rate": 0.0004914920725504518,
+ "loss": 2.1748604774475098,
+ "step": 1085
+ },
+ {
+ "epoch": 15.08362369337979,
+ "grad_norm": 0.07353878766298294,
+ "learning_rate": 0.000491285979372924,
+ "loss": 2.183143138885498,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097560975609756,
+ "grad_norm": 0.07123377919197083,
+ "learning_rate": 0.0004910797344967834,
+ "loss": 2.177582263946533,
+ "step": 1087
+ },
+ {
+ "epoch": 15.111498257839722,
+ "grad_norm": 0.0778692439198494,
+ "learning_rate": 0.000490873338088198,
+ "loss": 2.168792724609375,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125435540069686,
+ "grad_norm": 0.0788290873169899,
+ "learning_rate": 0.0004906667903134573,
+ "loss": 2.171933650970459,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139372822299652,
+ "grad_norm": 0.07082822173833847,
+ "learning_rate": 0.0004904600913389735,
+ "loss": 2.1847586631774902,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153310104529616,
+ "grad_norm": 0.07441762834787369,
+ "learning_rate": 0.0004902532413312799,
+ "loss": 2.187788963317871,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167247386759582,
+ "grad_norm": 0.07985713332891464,
+ "learning_rate": 0.0004900462404570321,
+ "loss": 2.182466506958008,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181184668989546,
+ "grad_norm": 0.07881174981594086,
+ "learning_rate": 0.0004898390888830067,
+ "loss": 2.189955711364746,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195121951219512,
+ "grad_norm": 0.07864490896463394,
+ "learning_rate": 0.0004896317867761022,
+ "loss": 2.191253185272217,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209059233449477,
+ "grad_norm": 0.07400403916835785,
+ "learning_rate": 0.000489424334303338,
+ "loss": 2.191585063934326,
+ "step": 1095
+ },
+ {
+ "epoch": 15.222996515679442,
+ "grad_norm": 0.06768042594194412,
+ "learning_rate": 0.0004892167316318548,
+ "loss": 2.184213638305664,
+ "step": 1096
+ },
+ {
+ "epoch": 15.236933797909408,
+ "grad_norm": 0.06592609733343124,
+ "learning_rate": 0.0004890089789289147,
+ "loss": 2.1969950199127197,
+ "step": 1097
+ },
+ {
+ "epoch": 15.250871080139373,
+ "grad_norm": 0.06763362139463425,
+ "learning_rate": 0.0004888010763618997,
+ "loss": 2.1938390731811523,
+ "step": 1098
+ },
+ {
+ "epoch": 15.264808362369338,
+ "grad_norm": 0.07065973430871964,
+ "learning_rate": 0.0004885930240983136,
+ "loss": 2.1912474632263184,
+ "step": 1099
+ },
+ {
+ "epoch": 15.278745644599303,
+ "grad_norm": 0.06374943256378174,
+ "learning_rate": 0.0004883848223057801,
+ "loss": 2.196803569793701,
+ "step": 1100
+ },
+ {
+ "epoch": 15.292682926829269,
+ "grad_norm": 0.06070345640182495,
+ "learning_rate": 0.00048817647115204386,
+ "loss": 2.185056209564209,
+ "step": 1101
+ },
+ {
+ "epoch": 15.306620209059233,
+ "grad_norm": 0.05391542240977287,
+ "learning_rate": 0.00048796797080496936,
+ "loss": 2.1775622367858887,
+ "step": 1102
+ },
+ {
+ "epoch": 15.320557491289199,
+ "grad_norm": 0.052196044474840164,
+ "learning_rate": 0.00048775932143254176,
+ "loss": 2.177192449569702,
+ "step": 1103
+ },
+ {
+ "epoch": 15.334494773519165,
+ "grad_norm": 0.05155641585588455,
+ "learning_rate": 0.0004875505232028661,
+ "loss": 2.1921792030334473,
+ "step": 1104
+ },
+ {
+ "epoch": 15.348432055749129,
+ "grad_norm": 0.049073997884988785,
+ "learning_rate": 0.0004873415762841672,
+ "loss": 2.184406042098999,
+ "step": 1105
+ },
+ {
+ "epoch": 15.362369337979095,
+ "grad_norm": 0.051046013832092285,
+ "learning_rate": 0.00048713248084478986,
+ "loss": 2.1981630325317383,
+ "step": 1106
+ },
+ {
+ "epoch": 15.376306620209059,
+ "grad_norm": 0.05237852409482002,
+ "learning_rate": 0.0004869232370531986,
+ "loss": 2.1774678230285645,
+ "step": 1107
+ },
+ {
+ "epoch": 15.390243902439025,
+ "grad_norm": 0.04945305734872818,
+ "learning_rate": 0.0004867138450779773,
+ "loss": 2.197721004486084,
+ "step": 1108
+ },
+ {
+ "epoch": 15.404181184668989,
+ "grad_norm": 0.04777121916413307,
+ "learning_rate": 0.0004865043050878293,
+ "loss": 2.197810649871826,
+ "step": 1109
+ },
+ {
+ "epoch": 15.418118466898955,
+ "grad_norm": 0.048481374979019165,
+ "learning_rate": 0.0004862946172515771,
+ "loss": 2.1903281211853027,
+ "step": 1110
+ },
+ {
+ "epoch": 15.43205574912892,
+ "grad_norm": 0.047138314694166183,
+ "learning_rate": 0.0004860847817381625,
+ "loss": 2.1983819007873535,
+ "step": 1111
+ },
+ {
+ "epoch": 15.445993031358885,
+ "grad_norm": 0.04804918169975281,
+ "learning_rate": 0.0004858747987166463,
+ "loss": 2.1968963146209717,
+ "step": 1112
+ },
+ {
+ "epoch": 15.45993031358885,
+ "grad_norm": 0.04860047250986099,
+ "learning_rate": 0.00048566466835620774,
+ "loss": 2.197075843811035,
+ "step": 1113
+ },
+ {
+ "epoch": 15.473867595818815,
+ "grad_norm": 0.047838736325502396,
+ "learning_rate": 0.0004854543908261453,
+ "loss": 2.1893868446350098,
+ "step": 1114
+ },
+ {
+ "epoch": 15.487804878048781,
+ "grad_norm": 0.04855247586965561,
+ "learning_rate": 0.0004852439662958756,
+ "loss": 2.2014851570129395,
+ "step": 1115
+ },
+ {
+ "epoch": 15.501742160278745,
+ "grad_norm": 0.04906660318374634,
+ "learning_rate": 0.000485033394934934,
+ "loss": 2.198460340499878,
+ "step": 1116
+ },
+ {
+ "epoch": 15.515679442508711,
+ "grad_norm": 0.049511708319187164,
+ "learning_rate": 0.000484822676912974,
+ "loss": 2.191267967224121,
+ "step": 1117
+ },
+ {
+ "epoch": 15.529616724738675,
+ "grad_norm": 0.047470301389694214,
+ "learning_rate": 0.0004846118123997674,
+ "loss": 2.1957900524139404,
+ "step": 1118
+ },
+ {
+ "epoch": 15.543554006968641,
+ "grad_norm": 0.04733480513095856,
+ "learning_rate": 0.00048440080156520387,
+ "loss": 2.20772123336792,
+ "step": 1119
+ },
+ {
+ "epoch": 15.557491289198607,
+ "grad_norm": 0.044856615364551544,
+ "learning_rate": 0.0004841896445792909,
+ "loss": 2.189664840698242,
+ "step": 1120
+ },
+ {
+ "epoch": 15.571428571428571,
+ "grad_norm": 0.047919657081365585,
+ "learning_rate": 0.000483978341612154,
+ "loss": 2.2144775390625,
+ "step": 1121
+ },
+ {
+ "epoch": 15.585365853658537,
+ "grad_norm": 0.04801390692591667,
+ "learning_rate": 0.0004837668928340362,
+ "loss": 2.186687469482422,
+ "step": 1122
+ },
+ {
+ "epoch": 15.599303135888501,
+ "grad_norm": 0.04735288396477699,
+ "learning_rate": 0.0004835552984152978,
+ "loss": 2.191199541091919,
+ "step": 1123
+ },
+ {
+ "epoch": 15.613240418118467,
+ "grad_norm": 0.0470828078687191,
+ "learning_rate": 0.00048334355852641664,
+ "loss": 2.205203056335449,
+ "step": 1124
+ },
+ {
+ "epoch": 15.627177700348431,
+ "grad_norm": 0.0449497289955616,
+ "learning_rate": 0.0004831316733379879,
+ "loss": 2.1997973918914795,
+ "step": 1125
+ },
+ {
+ "epoch": 15.641114982578397,
+ "grad_norm": 0.04837661236524582,
+ "learning_rate": 0.0004829196430207235,
+ "loss": 2.192153215408325,
+ "step": 1126
+ },
+ {
+ "epoch": 15.655052264808361,
+ "grad_norm": 0.04933372139930725,
+ "learning_rate": 0.00048270746774545237,
+ "loss": 2.190392255783081,
+ "step": 1127
+ },
+ {
+ "epoch": 15.668989547038327,
+ "grad_norm": 0.04986569285392761,
+ "learning_rate": 0.0004824951476831204,
+ "loss": 2.190096378326416,
+ "step": 1128
+ },
+ {
+ "epoch": 15.682926829268293,
+ "grad_norm": 0.04903094097971916,
+ "learning_rate": 0.00048228268300479015,
+ "loss": 2.2013378143310547,
+ "step": 1129
+ },
+ {
+ "epoch": 15.696864111498257,
+ "grad_norm": 0.04683993384242058,
+ "learning_rate": 0.0004820700738816403,
+ "loss": 2.1932640075683594,
+ "step": 1130
+ },
+ {
+ "epoch": 15.710801393728223,
+ "grad_norm": 0.04914448410272598,
+ "learning_rate": 0.0004818573204849664,
+ "loss": 2.1916818618774414,
+ "step": 1131
+ },
+ {
+ "epoch": 15.724738675958188,
+ "grad_norm": 0.0479729101061821,
+ "learning_rate": 0.00048164442298617987,
+ "loss": 2.1877894401550293,
+ "step": 1132
+ },
+ {
+ "epoch": 15.738675958188153,
+ "grad_norm": 0.04896317049860954,
+ "learning_rate": 0.00048143138155680844,
+ "loss": 2.200507164001465,
+ "step": 1133
+ },
+ {
+ "epoch": 15.752613240418118,
+ "grad_norm": 0.04853099212050438,
+ "learning_rate": 0.0004812181963684958,
+ "loss": 2.1982016563415527,
+ "step": 1134
+ },
+ {
+ "epoch": 15.766550522648084,
+ "grad_norm": 0.047448739409446716,
+ "learning_rate": 0.00048100486759300134,
+ "loss": 2.1995959281921387,
+ "step": 1135
+ },
+ {
+ "epoch": 15.78048780487805,
+ "grad_norm": 0.04720919579267502,
+ "learning_rate": 0.00048079139540220024,
+ "loss": 2.2148046493530273,
+ "step": 1136
+ },
+ {
+ "epoch": 15.794425087108014,
+ "grad_norm": 0.04254479706287384,
+ "learning_rate": 0.0004805777799680831,
+ "loss": 2.198126792907715,
+ "step": 1137
+ },
+ {
+ "epoch": 15.80836236933798,
+ "grad_norm": 0.0446142703294754,
+ "learning_rate": 0.0004803640214627561,
+ "loss": 2.2068254947662354,
+ "step": 1138
+ },
+ {
+ "epoch": 15.822299651567944,
+ "grad_norm": 0.044607460498809814,
+ "learning_rate": 0.0004801501200584406,
+ "loss": 2.1964797973632812,
+ "step": 1139
+ },
+ {
+ "epoch": 15.83623693379791,
+ "grad_norm": 0.04692309722304344,
+ "learning_rate": 0.000479936075927473,
+ "loss": 2.2058191299438477,
+ "step": 1140
+ },
+ {
+ "epoch": 15.850174216027874,
+ "grad_norm": 0.04527280852198601,
+ "learning_rate": 0.000479721889242305,
+ "loss": 2.2034077644348145,
+ "step": 1141
+ },
+ {
+ "epoch": 15.86411149825784,
+ "grad_norm": 0.044546857476234436,
+ "learning_rate": 0.0004795075601755027,
+ "loss": 2.199248790740967,
+ "step": 1142
+ },
+ {
+ "epoch": 15.878048780487806,
+ "grad_norm": 0.044123925268650055,
+ "learning_rate": 0.00047929308889974737,
+ "loss": 2.1891608238220215,
+ "step": 1143
+ },
+ {
+ "epoch": 15.89198606271777,
+ "grad_norm": 0.04888012632727623,
+ "learning_rate": 0.0004790784755878346,
+ "loss": 2.1975278854370117,
+ "step": 1144
+ },
+ {
+ "epoch": 15.905923344947736,
+ "grad_norm": 0.0479818694293499,
+ "learning_rate": 0.00047886372041267434,
+ "loss": 2.2012686729431152,
+ "step": 1145
+ },
+ {
+ "epoch": 15.9198606271777,
+ "grad_norm": 0.04637312889099121,
+ "learning_rate": 0.0004786488235472909,
+ "loss": 2.2082278728485107,
+ "step": 1146
+ },
+ {
+ "epoch": 15.933797909407666,
+ "grad_norm": 0.04439574480056763,
+ "learning_rate": 0.0004784337851648231,
+ "loss": 2.1968092918395996,
+ "step": 1147
+ },
+ {
+ "epoch": 15.94773519163763,
+ "grad_norm": 0.04690313711762428,
+ "learning_rate": 0.0004782186054385233,
+ "loss": 2.207057237625122,
+ "step": 1148
+ },
+ {
+ "epoch": 15.961672473867596,
+ "grad_norm": 0.04463008791208267,
+ "learning_rate": 0.0004780032845417579,
+ "loss": 2.204643487930298,
+ "step": 1149
+ },
+ {
+ "epoch": 15.975609756097562,
+ "grad_norm": 0.045115210115909576,
+ "learning_rate": 0.000477787822648007,
+ "loss": 2.186087131500244,
+ "step": 1150
+ },
+ {
+ "epoch": 15.989547038327526,
+ "grad_norm": 0.048475850373506546,
+ "learning_rate": 0.00047757221993086443,
+ "loss": 2.172334671020508,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.03669014573097229,
+ "learning_rate": 0.00047735647656403754,
+ "loss": 1.6437406539916992,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 0.6166133284568787,
+ "eval_runtime": 41.0119,
+ "eval_samples_per_second": 59.544,
+ "eval_steps_per_second": 0.488,
+ "step": 1152
+ },
+ {
+ "epoch": 16.013937282229964,
+ "grad_norm": 0.04941252991557121,
+ "learning_rate": 0.0004771405927213467,
+ "loss": 2.1503076553344727,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027874564459932,
+ "grad_norm": 0.06683871150016785,
+ "learning_rate": 0.0004769245685767255,
+ "loss": 2.1464710235595703,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041811846689896,
+ "grad_norm": 0.0639740377664566,
+ "learning_rate": 0.00047670840430422113,
+ "loss": 2.1400065422058105,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05574912891986,
+ "grad_norm": 0.06087620183825493,
+ "learning_rate": 0.0004764921000779928,
+ "loss": 2.1587162017822266,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069686411149824,
+ "grad_norm": 0.062141239643096924,
+ "learning_rate": 0.0004762756560723131,
+ "loss": 2.1633496284484863,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083623693379792,
+ "grad_norm": 0.06054656580090523,
+ "learning_rate": 0.0004760590724615671,
+ "loss": 2.140836477279663,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097560975609756,
+ "grad_norm": 0.06489263474941254,
+ "learning_rate": 0.0004758423494202522,
+ "loss": 2.1461167335510254,
+ "step": 1159
+ },
+ {
+ "epoch": 16.11149825783972,
+ "grad_norm": 0.07237063348293304,
+ "learning_rate": 0.0004756254871229782,
+ "loss": 2.143646240234375,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125435540069688,
+ "grad_norm": 0.06934516131877899,
+ "learning_rate": 0.00047540848574446724,
+ "loss": 2.147749900817871,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139372822299652,
+ "grad_norm": 0.06326598674058914,
+ "learning_rate": 0.0004751913454595535,
+ "loss": 2.1485419273376465,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153310104529616,
+ "grad_norm": 0.05871129781007767,
+ "learning_rate": 0.0004749740664431828,
+ "loss": 2.1547937393188477,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16724738675958,
+ "grad_norm": 0.058887775987386703,
+ "learning_rate": 0.000474756648870413,
+ "loss": 2.154172420501709,
+ "step": 1164
+ },
+ {
+ "epoch": 16.181184668989548,
+ "grad_norm": 0.06326179951429367,
+ "learning_rate": 0.0004745390929164134,
+ "loss": 2.159712314605713,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195121951219512,
+ "grad_norm": 0.061434682458639145,
+ "learning_rate": 0.0004743213987564651,
+ "loss": 2.1372804641723633,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209059233449477,
+ "grad_norm": 0.06278357654809952,
+ "learning_rate": 0.0004741035665659603,
+ "loss": 2.1623623371124268,
+ "step": 1167
+ },
+ {
+ "epoch": 16.222996515679444,
+ "grad_norm": 0.06233280152082443,
+ "learning_rate": 0.00047388559652040236,
+ "loss": 2.150796890258789,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23693379790941,
+ "grad_norm": 0.0610053725540638,
+ "learning_rate": 0.0004736674887954059,
+ "loss": 2.152132987976074,
+ "step": 1169
+ },
+ {
+ "epoch": 16.250871080139373,
+ "grad_norm": 0.06266728788614273,
+ "learning_rate": 0.0004734492435666963,
+ "loss": 2.1564221382141113,
+ "step": 1170
+ },
+ {
+ "epoch": 16.264808362369337,
+ "grad_norm": 0.0677572563290596,
+ "learning_rate": 0.0004732308610101099,
+ "loss": 2.160862445831299,
+ "step": 1171
+ },
+ {
+ "epoch": 16.278745644599304,
+ "grad_norm": 0.06443636119365692,
+ "learning_rate": 0.00047301234130159346,
+ "loss": 2.164417266845703,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29268292682927,
+ "grad_norm": 0.06273599714040756,
+ "learning_rate": 0.00047279368461720445,
+ "loss": 2.1794919967651367,
+ "step": 1173
+ },
+ {
+ "epoch": 16.306620209059233,
+ "grad_norm": 0.05729149281978607,
+ "learning_rate": 0.0004725748911331106,
+ "loss": 2.170616626739502,
+ "step": 1174
+ },
+ {
+ "epoch": 16.320557491289197,
+ "grad_norm": 0.05595673620700836,
+ "learning_rate": 0.0004723559610255898,
+ "loss": 2.156553268432617,
+ "step": 1175
+ },
+ {
+ "epoch": 16.334494773519165,
+ "grad_norm": 0.05238134786486626,
+ "learning_rate": 0.0004721368944710302,
+ "loss": 2.174781322479248,
+ "step": 1176
+ },
+ {
+ "epoch": 16.34843205574913,
+ "grad_norm": 0.05605665221810341,
+ "learning_rate": 0.00047191769164592974,
+ "loss": 2.1708836555480957,
+ "step": 1177
+ },
+ {
+ "epoch": 16.362369337979093,
+ "grad_norm": 0.058390554040670395,
+ "learning_rate": 0.000471698352726896,
+ "loss": 2.166501998901367,
+ "step": 1178
+ },
+ {
+ "epoch": 16.37630662020906,
+ "grad_norm": 0.05810759216547012,
+ "learning_rate": 0.0004714788778906467,
+ "loss": 2.162776470184326,
+ "step": 1179
+ },
+ {
+ "epoch": 16.390243902439025,
+ "grad_norm": 0.053019024431705475,
+ "learning_rate": 0.0004712592673140084,
+ "loss": 2.170766830444336,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40418118466899,
+ "grad_norm": 0.051433514803647995,
+ "learning_rate": 0.00047103952117391764,
+ "loss": 2.156322479248047,
+ "step": 1181
+ },
+ {
+ "epoch": 16.418118466898953,
+ "grad_norm": 0.05271990969777107,
+ "learning_rate": 0.00047081963964741986,
+ "loss": 2.1634387969970703,
+ "step": 1182
+ },
+ {
+ "epoch": 16.43205574912892,
+ "grad_norm": 0.048728011548519135,
+ "learning_rate": 0.0004705996229116696,
+ "loss": 2.1454968452453613,
+ "step": 1183
+ },
+ {
+ "epoch": 16.445993031358885,
+ "grad_norm": 0.04880265146493912,
+ "learning_rate": 0.0004703794711439304,
+ "loss": 2.164184093475342,
+ "step": 1184
+ },
+ {
+ "epoch": 16.45993031358885,
+ "grad_norm": 0.05487549304962158,
+ "learning_rate": 0.0004701591845215744,
+ "loss": 2.1683168411254883,
+ "step": 1185
+ },
+ {
+ "epoch": 16.473867595818817,
+ "grad_norm": 0.05033471807837486,
+ "learning_rate": 0.00046993876322208274,
+ "loss": 2.178168296813965,
+ "step": 1186
+ },
+ {
+ "epoch": 16.48780487804878,
+ "grad_norm": 0.0512198805809021,
+ "learning_rate": 0.0004697182074230448,
+ "loss": 2.1766600608825684,
+ "step": 1187
+ },
+ {
+ "epoch": 16.501742160278745,
+ "grad_norm": 0.050174590200185776,
+ "learning_rate": 0.0004694975173021586,
+ "loss": 2.177060604095459,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51567944250871,
+ "grad_norm": 0.050394315272569656,
+ "learning_rate": 0.0004692766930372299,
+ "loss": 2.1681301593780518,
+ "step": 1189
+ },
+ {
+ "epoch": 16.529616724738677,
+ "grad_norm": 0.05456200987100601,
+ "learning_rate": 0.00046905573480617285,
+ "loss": 2.1732399463653564,
+ "step": 1190
+ },
+ {
+ "epoch": 16.54355400696864,
+ "grad_norm": 0.05683460086584091,
+ "learning_rate": 0.00046883464278700966,
+ "loss": 2.1721444129943848,
+ "step": 1191
+ },
+ {
+ "epoch": 16.557491289198605,
+ "grad_norm": 0.054592303931713104,
+ "learning_rate": 0.0004686134171578702,
+ "loss": 2.1711673736572266,
+ "step": 1192
+ },
+ {
+ "epoch": 16.571428571428573,
+ "grad_norm": 0.04853744059801102,
+ "learning_rate": 0.00046839205809699183,
+ "loss": 2.1945948600769043,
+ "step": 1193
+ },
+ {
+ "epoch": 16.585365853658537,
+ "grad_norm": 0.04630011320114136,
+ "learning_rate": 0.0004681705657827195,
+ "loss": 2.158024311065674,
+ "step": 1194
+ },
+ {
+ "epoch": 16.5993031358885,
+ "grad_norm": 0.050844110548496246,
+ "learning_rate": 0.0004679489403935058,
+ "loss": 2.1652145385742188,
+ "step": 1195
+ },
+ {
+ "epoch": 16.613240418118465,
+ "grad_norm": 0.047757863998413086,
+ "learning_rate": 0.00046772718210791005,
+ "loss": 2.183868408203125,
+ "step": 1196
+ },
+ {
+ "epoch": 16.627177700348433,
+ "grad_norm": 0.045895107090473175,
+ "learning_rate": 0.00046750529110459903,
+ "loss": 2.1651611328125,
+ "step": 1197
+ },
+ {
+ "epoch": 16.641114982578397,
+ "grad_norm": 0.0506361722946167,
+ "learning_rate": 0.0004672832675623463,
+ "loss": 2.175286293029785,
+ "step": 1198
+ },
+ {
+ "epoch": 16.65505226480836,
+ "grad_norm": 0.049629244953393936,
+ "learning_rate": 0.0004670611116600322,
+ "loss": 2.1572513580322266,
+ "step": 1199
+ },
+ {
+ "epoch": 16.66898954703833,
+ "grad_norm": 0.04853643849492073,
+ "learning_rate": 0.00046683882357664375,
+ "loss": 2.1820926666259766,
+ "step": 1200
+ },
+ {
+ "epoch": 16.682926829268293,
+ "grad_norm": 0.0491788275539875,
+ "learning_rate": 0.0004666164034912744,
+ "loss": 2.1798288822174072,
+ "step": 1201
+ },
+ {
+ "epoch": 16.696864111498257,
+ "grad_norm": 0.050360891968011856,
+ "learning_rate": 0.00046639385158312406,
+ "loss": 2.159900426864624,
+ "step": 1202
+ },
+ {
+ "epoch": 16.71080139372822,
+ "grad_norm": 0.052317969501018524,
+ "learning_rate": 0.00046617116803149874,
+ "loss": 2.188831329345703,
+ "step": 1203
+ },
+ {
+ "epoch": 16.72473867595819,
+ "grad_norm": 0.05087872967123985,
+ "learning_rate": 0.00046594835301581053,
+ "loss": 2.1665854454040527,
+ "step": 1204
+ },
+ {
+ "epoch": 16.738675958188153,
+ "grad_norm": 0.05026745796203613,
+ "learning_rate": 0.00046572540671557745,
+ "loss": 2.1804490089416504,
+ "step": 1205
+ },
+ {
+ "epoch": 16.752613240418118,
+ "grad_norm": 0.05136656388640404,
+ "learning_rate": 0.00046550232931042346,
+ "loss": 2.1803054809570312,
+ "step": 1206
+ },
+ {
+ "epoch": 16.766550522648085,
+ "grad_norm": 0.04974747449159622,
+ "learning_rate": 0.00046527912098007787,
+ "loss": 2.1852850914001465,
+ "step": 1207
+ },
+ {
+ "epoch": 16.78048780487805,
+ "grad_norm": 0.048434171825647354,
+ "learning_rate": 0.00046505578190437564,
+ "loss": 2.179105043411255,
+ "step": 1208
+ },
+ {
+ "epoch": 16.794425087108014,
+ "grad_norm": 0.052139151841402054,
+ "learning_rate": 0.0004648323122632569,
+ "loss": 2.1896872520446777,
+ "step": 1209
+ },
+ {
+ "epoch": 16.808362369337978,
+ "grad_norm": 0.0564812608063221,
+ "learning_rate": 0.0004646087122367673,
+ "loss": 2.1947436332702637,
+ "step": 1210
+ },
+ {
+ "epoch": 16.822299651567945,
+ "grad_norm": 0.05525265261530876,
+ "learning_rate": 0.00046438498200505727,
+ "loss": 2.1992716789245605,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83623693379791,
+ "grad_norm": 0.053308531641960144,
+ "learning_rate": 0.0004641611217483821,
+ "loss": 2.168307065963745,
+ "step": 1212
+ },
+ {
+ "epoch": 16.850174216027874,
+ "grad_norm": 0.053110528737306595,
+ "learning_rate": 0.00046393713164710217,
+ "loss": 2.1733713150024414,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86411149825784,
+ "grad_norm": 0.05017722025513649,
+ "learning_rate": 0.00046371301188168204,
+ "loss": 2.1804280281066895,
+ "step": 1214
+ },
+ {
+ "epoch": 16.878048780487806,
+ "grad_norm": 0.05085223168134689,
+ "learning_rate": 0.0004634887626326911,
+ "loss": 2.1635611057281494,
+ "step": 1215
+ },
+ {
+ "epoch": 16.89198606271777,
+ "grad_norm": 0.04969240725040436,
+ "learning_rate": 0.00046326438408080293,
+ "loss": 2.183300495147705,
+ "step": 1216
+ },
+ {
+ "epoch": 16.905923344947734,
+ "grad_norm": 0.05200999230146408,
+ "learning_rate": 0.00046303987640679517,
+ "loss": 2.1979541778564453,
+ "step": 1217
+ },
+ {
+ "epoch": 16.9198606271777,
+ "grad_norm": 0.04787272959947586,
+ "learning_rate": 0.0004628152397915498,
+ "loss": 2.1646995544433594,
+ "step": 1218
+ },
+ {
+ "epoch": 16.933797909407666,
+ "grad_norm": 0.04960496723651886,
+ "learning_rate": 0.00046259047441605215,
+ "loss": 2.1800191402435303,
+ "step": 1219
+ },
+ {
+ "epoch": 16.94773519163763,
+ "grad_norm": 0.049237579107284546,
+ "learning_rate": 0.0004623655804613919,
+ "loss": 2.182941436767578,
+ "step": 1220
+ },
+ {
+ "epoch": 16.961672473867594,
+ "grad_norm": 0.04768526926636696,
+ "learning_rate": 0.00046214055810876194,
+ "loss": 2.179150104522705,
+ "step": 1221
+ },
+ {
+ "epoch": 16.975609756097562,
+ "grad_norm": 0.04873532056808472,
+ "learning_rate": 0.00046191540753945886,
+ "loss": 2.1794910430908203,
+ "step": 1222
+ },
+ {
+ "epoch": 16.989547038327526,
+ "grad_norm": 0.04799315333366394,
+ "learning_rate": 0.00046169012893488204,
+ "loss": 2.1740200519561768,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.040487222373485565,
+ "learning_rate": 0.00046146472247653484,
+ "loss": 1.643998622894287,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 0.6177874207496643,
+ "eval_runtime": 39.9799,
+ "eval_samples_per_second": 61.081,
+ "eval_steps_per_second": 0.5,
+ "step": 1224
+ },
+ {
+ "epoch": 17.013937282229964,
+ "grad_norm": 0.05245300009846687,
+ "learning_rate": 0.0004612391883460228,
+ "loss": 2.138477087020874,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027874564459932,
+ "grad_norm": 0.06304061412811279,
+ "learning_rate": 0.00046101352672505493,
+ "loss": 2.128952741622925,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041811846689896,
+ "grad_norm": 0.0575832761824131,
+ "learning_rate": 0.0004607877377954426,
+ "loss": 2.134880542755127,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05574912891986,
+ "grad_norm": 0.06044311448931694,
+ "learning_rate": 0.00046056182173909995,
+ "loss": 2.121551513671875,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069686411149824,
+ "grad_norm": 0.058043431490659714,
+ "learning_rate": 0.0004603357787380435,
+ "loss": 2.127577781677246,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083623693379792,
+ "grad_norm": 0.05703338608145714,
+ "learning_rate": 0.0004601096089743919,
+ "loss": 2.1422252655029297,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097560975609756,
+ "grad_norm": 0.05726749077439308,
+ "learning_rate": 0.0004598833126303661,
+ "loss": 2.114379644393921,
+ "step": 1231
+ },
+ {
+ "epoch": 17.11149825783972,
+ "grad_norm": 0.05507425218820572,
+ "learning_rate": 0.00045965688988828884,
+ "loss": 2.1125502586364746,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125435540069688,
+ "grad_norm": 0.058553360402584076,
+ "learning_rate": 0.00045943034093058507,
+ "loss": 2.1171164512634277,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139372822299652,
+ "grad_norm": 0.06292160600423813,
+ "learning_rate": 0.000459203665939781,
+ "loss": 2.114016532897949,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153310104529616,
+ "grad_norm": 0.06509732455015182,
+ "learning_rate": 0.00045897686509850487,
+ "loss": 2.1301136016845703,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16724738675958,
+ "grad_norm": 0.06321847438812256,
+ "learning_rate": 0.0004587499385894856,
+ "loss": 2.128077983856201,
+ "step": 1236
+ },
+ {
+ "epoch": 17.181184668989548,
+ "grad_norm": 0.062314026057720184,
+ "learning_rate": 0.000458522886595554,
+ "loss": 2.139747142791748,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195121951219512,
+ "grad_norm": 0.07283712178468704,
+ "learning_rate": 0.0004582957092996418,
+ "loss": 2.1305007934570312,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209059233449477,
+ "grad_norm": 0.08525235950946808,
+ "learning_rate": 0.0004580684068847817,
+ "loss": 2.126866340637207,
+ "step": 1239
+ },
+ {
+ "epoch": 17.222996515679444,
+ "grad_norm": 0.08363549411296844,
+ "learning_rate": 0.0004578409795341069,
+ "loss": 2.1400678157806396,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23693379790941,
+ "grad_norm": 0.06837999820709229,
+ "learning_rate": 0.0004576134274308517,
+ "loss": 2.1413705348968506,
+ "step": 1241
+ },
+ {
+ "epoch": 17.250871080139373,
+ "grad_norm": 0.07704652845859528,
+ "learning_rate": 0.0004573857507583506,
+ "loss": 2.148033857345581,
+ "step": 1242
+ },
+ {
+ "epoch": 17.264808362369337,
+ "grad_norm": 0.07117756456136703,
+ "learning_rate": 0.00045715794970003874,
+ "loss": 2.131544828414917,
+ "step": 1243
+ },
+ {
+ "epoch": 17.278745644599304,
+ "grad_norm": 0.06029726192355156,
+ "learning_rate": 0.0004569300244394513,
+ "loss": 2.139054775238037,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29268292682927,
+ "grad_norm": 0.06284599751234055,
+ "learning_rate": 0.0004567019751602234,
+ "loss": 2.153508186340332,
+ "step": 1245
+ },
+ {
+ "epoch": 17.306620209059233,
+ "grad_norm": 0.06004280224442482,
+ "learning_rate": 0.00045647380204609034,
+ "loss": 2.1395351886749268,
+ "step": 1246
+ },
+ {
+ "epoch": 17.320557491289197,
+ "grad_norm": 0.05850320681929588,
+ "learning_rate": 0.00045624550528088704,
+ "loss": 2.1343579292297363,
+ "step": 1247
+ },
+ {
+ "epoch": 17.334494773519165,
+ "grad_norm": 0.05615168437361717,
+ "learning_rate": 0.00045601708504854803,
+ "loss": 2.1352741718292236,
+ "step": 1248
+ },
+ {
+ "epoch": 17.34843205574913,
+ "grad_norm": 0.060361411422491074,
+ "learning_rate": 0.0004557885415331074,
+ "loss": 2.1650805473327637,
+ "step": 1249
+ },
+ {
+ "epoch": 17.362369337979093,
+ "grad_norm": 0.057178955525159836,
+ "learning_rate": 0.00045555987491869865,
+ "loss": 2.1436054706573486,
+ "step": 1250
+ },
+ {
+ "epoch": 17.37630662020906,
+ "grad_norm": 0.06044763699173927,
+ "learning_rate": 0.00045533108538955416,
+ "loss": 2.1491849422454834,
+ "step": 1251
+ },
+ {
+ "epoch": 17.390243902439025,
+ "grad_norm": 0.061660051345825195,
+ "learning_rate": 0.0004551021731300055,
+ "loss": 2.1438241004943848,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40418118466899,
+ "grad_norm": 0.05951886624097824,
+ "learning_rate": 0.0004548731383244834,
+ "loss": 2.1414897441864014,
+ "step": 1253
+ },
+ {
+ "epoch": 17.418118466898953,
+ "grad_norm": 0.05718380957841873,
+ "learning_rate": 0.0004546439811575169,
+ "loss": 2.139619827270508,
+ "step": 1254
+ },
+ {
+ "epoch": 17.43205574912892,
+ "grad_norm": 0.0523972325026989,
+ "learning_rate": 0.00045441470181373375,
+ "loss": 2.131605625152588,
+ "step": 1255
+ },
+ {
+ "epoch": 17.445993031358885,
+ "grad_norm": 0.05557102710008621,
+ "learning_rate": 0.0004541853004778603,
+ "loss": 2.151350259780884,
+ "step": 1256
+ },
+ {
+ "epoch": 17.45993031358885,
+ "grad_norm": 0.053447823971509933,
+ "learning_rate": 0.00045395577733472087,
+ "loss": 2.1357831954956055,
+ "step": 1257
+ },
+ {
+ "epoch": 17.473867595818817,
+ "grad_norm": 0.05254426226019859,
+ "learning_rate": 0.0004537261325692383,
+ "loss": 2.147961139678955,
+ "step": 1258
+ },
+ {
+ "epoch": 17.48780487804878,
+ "grad_norm": 0.052484504878520966,
+ "learning_rate": 0.0004534963663664332,
+ "loss": 2.1512255668640137,
+ "step": 1259
+ },
+ {
+ "epoch": 17.501742160278745,
+ "grad_norm": 0.05347288027405739,
+ "learning_rate": 0.00045326647891142394,
+ "loss": 2.15310001373291,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51567944250871,
+ "grad_norm": 0.05540700629353523,
+ "learning_rate": 0.0004530364703894268,
+ "loss": 2.1427979469299316,
+ "step": 1261
+ },
+ {
+ "epoch": 17.529616724738677,
+ "grad_norm": 0.05409472435712814,
+ "learning_rate": 0.0004528063409857554,
+ "loss": 2.17159366607666,
+ "step": 1262
+ },
+ {
+ "epoch": 17.54355400696864,
+ "grad_norm": 0.049905259162187576,
+ "learning_rate": 0.0004525760908858209,
+ "loss": 2.1460530757904053,
+ "step": 1263
+ },
+ {
+ "epoch": 17.557491289198605,
+ "grad_norm": 0.052097104489803314,
+ "learning_rate": 0.00045234572027513174,
+ "loss": 2.155837059020996,
+ "step": 1264
+ },
+ {
+ "epoch": 17.571428571428573,
+ "grad_norm": 0.05186805874109268,
+ "learning_rate": 0.00045211522933929317,
+ "loss": 2.150564193725586,
+ "step": 1265
+ },
+ {
+ "epoch": 17.585365853658537,
+ "grad_norm": 0.05186301842331886,
+ "learning_rate": 0.0004518846182640077,
+ "loss": 2.1461782455444336,
+ "step": 1266
+ },
+ {
+ "epoch": 17.5993031358885,
+ "grad_norm": 0.05042846128344536,
+ "learning_rate": 0.0004516538872350744,
+ "loss": 2.159538507461548,
+ "step": 1267
+ },
+ {
+ "epoch": 17.613240418118465,
+ "grad_norm": 0.05357243865728378,
+ "learning_rate": 0.0004514230364383893,
+ "loss": 2.1775975227355957,
+ "step": 1268
+ },
+ {
+ "epoch": 17.627177700348433,
+ "grad_norm": 0.05251279100775719,
+ "learning_rate": 0.0004511920660599444,
+ "loss": 2.1498348712921143,
+ "step": 1269
+ },
+ {
+ "epoch": 17.641114982578397,
+ "grad_norm": 0.05563712120056152,
+ "learning_rate": 0.0004509609762858287,
+ "loss": 2.1542813777923584,
+ "step": 1270
+ },
+ {
+ "epoch": 17.65505226480836,
+ "grad_norm": 0.05359400436282158,
+ "learning_rate": 0.0004507297673022269,
+ "loss": 2.150785446166992,
+ "step": 1271
+ },
+ {
+ "epoch": 17.66898954703833,
+ "grad_norm": 0.05313696339726448,
+ "learning_rate": 0.0004504984392954199,
+ "loss": 2.148235321044922,
+ "step": 1272
+ },
+ {
+ "epoch": 17.682926829268293,
+ "grad_norm": 0.05827540531754494,
+ "learning_rate": 0.00045026699245178444,
+ "loss": 2.151808738708496,
+ "step": 1273
+ },
+ {
+ "epoch": 17.696864111498257,
+ "grad_norm": 0.05343877896666527,
+ "learning_rate": 0.0004500354269577932,
+ "loss": 2.1618194580078125,
+ "step": 1274
+ },
+ {
+ "epoch": 17.71080139372822,
+ "grad_norm": 0.05064912885427475,
+ "learning_rate": 0.00044980374300001414,
+ "loss": 2.148162841796875,
+ "step": 1275
+ },
+ {
+ "epoch": 17.72473867595819,
+ "grad_norm": 0.05388092249631882,
+ "learning_rate": 0.00044957194076511095,
+ "loss": 2.1632771492004395,
+ "step": 1276
+ },
+ {
+ "epoch": 17.738675958188153,
+ "grad_norm": 0.05264539271593094,
+ "learning_rate": 0.0004493400204398426,
+ "loss": 2.160672187805176,
+ "step": 1277
+ },
+ {
+ "epoch": 17.752613240418118,
+ "grad_norm": 0.052000492811203,
+ "learning_rate": 0.00044910798221106283,
+ "loss": 2.1609997749328613,
+ "step": 1278
+ },
+ {
+ "epoch": 17.766550522648085,
+ "grad_norm": 0.05338827148079872,
+ "learning_rate": 0.00044887582626572105,
+ "loss": 2.15794038772583,
+ "step": 1279
+ },
+ {
+ "epoch": 17.78048780487805,
+ "grad_norm": 0.05201420187950134,
+ "learning_rate": 0.0004486435527908606,
+ "loss": 2.153334856033325,
+ "step": 1280
+ },
+ {
+ "epoch": 17.794425087108014,
+ "grad_norm": 0.05256861820816994,
+ "learning_rate": 0.0004484111619736205,
+ "loss": 2.16025972366333,
+ "step": 1281
+ },
+ {
+ "epoch": 17.808362369337978,
+ "grad_norm": 0.050007306039333344,
+ "learning_rate": 0.0004481786540012336,
+ "loss": 2.1449337005615234,
+ "step": 1282
+ },
+ {
+ "epoch": 17.822299651567945,
+ "grad_norm": 0.05070459097623825,
+ "learning_rate": 0.00044794602906102747,
+ "loss": 2.169527292251587,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83623693379791,
+ "grad_norm": 0.05200141668319702,
+ "learning_rate": 0.00044771328734042367,
+ "loss": 2.163435935974121,
+ "step": 1284
+ },
+ {
+ "epoch": 17.850174216027874,
+ "grad_norm": 0.04956226050853729,
+ "learning_rate": 0.00044748042902693815,
+ "loss": 2.1637909412384033,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86411149825784,
+ "grad_norm": 0.05235756188631058,
+ "learning_rate": 0.0004472474543081805,
+ "loss": 2.1584506034851074,
+ "step": 1286
+ },
+ {
+ "epoch": 17.878048780487806,
+ "grad_norm": 0.05084889009594917,
+ "learning_rate": 0.00044701436337185444,
+ "loss": 2.146690845489502,
+ "step": 1287
+ },
+ {
+ "epoch": 17.89198606271777,
+ "grad_norm": 0.05055645480751991,
+ "learning_rate": 0.00044678115640575704,
+ "loss": 2.1586077213287354,
+ "step": 1288
+ },
+ {
+ "epoch": 17.905923344947734,
+ "grad_norm": 0.05491330847144127,
+ "learning_rate": 0.0004465478335977788,
+ "loss": 2.149641752243042,
+ "step": 1289
+ },
+ {
+ "epoch": 17.9198606271777,
+ "grad_norm": 0.05221051350235939,
+ "learning_rate": 0.00044631439513590393,
+ "loss": 2.1694397926330566,
+ "step": 1290
+ },
+ {
+ "epoch": 17.933797909407666,
+ "grad_norm": 0.04915725812315941,
+ "learning_rate": 0.00044608084120820933,
+ "loss": 2.145416736602783,
+ "step": 1291
+ },
+ {
+ "epoch": 17.94773519163763,
+ "grad_norm": 0.04936422407627106,
+ "learning_rate": 0.0004458471720028655,
+ "loss": 2.1609673500061035,
+ "step": 1292
+ },
+ {
+ "epoch": 17.961672473867594,
+ "grad_norm": 0.0533466637134552,
+ "learning_rate": 0.0004456133877081352,
+ "loss": 2.162301540374756,
+ "step": 1293
+ },
+ {
+ "epoch": 17.975609756097562,
+ "grad_norm": 0.05231599137187004,
+ "learning_rate": 0.0004453794885123744,
+ "loss": 2.1669540405273438,
+ "step": 1294
+ },
+ {
+ "epoch": 17.989547038327526,
+ "grad_norm": 0.05081851780414581,
+ "learning_rate": 0.00044514547460403135,
+ "loss": 2.15706205368042,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.04242361709475517,
+ "learning_rate": 0.00044491134617164694,
+ "loss": 1.6120411157608032,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 0.6197657585144043,
+ "eval_runtime": 40.1015,
+ "eval_samples_per_second": 60.895,
+ "eval_steps_per_second": 0.499,
+ "step": 1296
+ },
+ {
+ "epoch": 18.013937282229964,
+ "grad_norm": 0.04986736550927162,
+ "learning_rate": 0.000444677103403854,
+ "loss": 2.1041247844696045,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027874564459932,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0004444427464893779,
+ "loss": 2.118490219116211,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041811846689896,
+ "grad_norm": 0.06156685948371887,
+ "learning_rate": 0.00044420827561703573,
+ "loss": 2.1154561042785645,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05574912891986,
+ "grad_norm": 0.06687360256910324,
+ "learning_rate": 0.0004439736909757365,
+ "loss": 2.103177547454834,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069686411149824,
+ "grad_norm": 0.06271911412477493,
+ "learning_rate": 0.0004437389927544805,
+ "loss": 2.0895836353302,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083623693379792,
+ "grad_norm": 0.059448983520269394,
+ "learning_rate": 0.0004435041811423603,
+ "loss": 2.121011734008789,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097560975609756,
+ "grad_norm": 0.06396909803152084,
+ "learning_rate": 0.0004432692563285591,
+ "loss": 2.117781162261963,
+ "step": 1303
+ },
+ {
+ "epoch": 18.11149825783972,
+ "grad_norm": 0.07109276950359344,
+ "learning_rate": 0.0004430342185023517,
+ "loss": 2.1139042377471924,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125435540069688,
+ "grad_norm": 0.0748843252658844,
+ "learning_rate": 0.0004427990678531038,
+ "loss": 2.1079554557800293,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139372822299652,
+ "grad_norm": 0.06985905021429062,
+ "learning_rate": 0.0004425638045702719,
+ "loss": 2.0937323570251465,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153310104529616,
+ "grad_norm": 0.06744816899299622,
+ "learning_rate": 0.0004423284288434035,
+ "loss": 2.1189656257629395,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16724738675958,
+ "grad_norm": 0.06989820301532745,
+ "learning_rate": 0.00044209294086213657,
+ "loss": 2.1067135334014893,
+ "step": 1308
+ },
+ {
+ "epoch": 18.181184668989548,
+ "grad_norm": 0.06230093166232109,
+ "learning_rate": 0.00044185734081619957,
+ "loss": 2.105956554412842,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195121951219512,
+ "grad_norm": 0.05950392782688141,
+ "learning_rate": 0.000441621628895411,
+ "loss": 2.1188323497772217,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209059233449477,
+ "grad_norm": 0.058130305260419846,
+ "learning_rate": 0.00044138580528967986,
+ "loss": 2.129612445831299,
+ "step": 1311
+ },
+ {
+ "epoch": 18.222996515679444,
+ "grad_norm": 0.05574113503098488,
+ "learning_rate": 0.0004411498701890051,
+ "loss": 2.106229782104492,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23693379790941,
+ "grad_norm": 0.059406183660030365,
+ "learning_rate": 0.0004409138237834751,
+ "loss": 2.1047778129577637,
+ "step": 1313
+ },
+ {
+ "epoch": 18.250871080139373,
+ "grad_norm": 0.05741738900542259,
+ "learning_rate": 0.00044067766626326835,
+ "loss": 2.1045103073120117,
+ "step": 1314
+ },
+ {
+ "epoch": 18.264808362369337,
+ "grad_norm": 0.05358671396970749,
+ "learning_rate": 0.00044044139781865255,
+ "loss": 2.1150290966033936,
+ "step": 1315
+ },
+ {
+ "epoch": 18.278745644599304,
+ "grad_norm": 0.054804909974336624,
+ "learning_rate": 0.00044020501863998516,
+ "loss": 2.106473922729492,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29268292682927,
+ "grad_norm": 0.053977105766534805,
+ "learning_rate": 0.00043996852891771236,
+ "loss": 2.1170592308044434,
+ "step": 1317
+ },
+ {
+ "epoch": 18.306620209059233,
+ "grad_norm": 0.055874355137348175,
+ "learning_rate": 0.00043973192884236983,
+ "loss": 2.116635322570801,
+ "step": 1318
+ },
+ {
+ "epoch": 18.320557491289197,
+ "grad_norm": 0.05919129028916359,
+ "learning_rate": 0.0004394952186045819,
+ "loss": 2.1295974254608154,
+ "step": 1319
+ },
+ {
+ "epoch": 18.334494773519165,
+ "grad_norm": 0.05559816211462021,
+ "learning_rate": 0.00043925839839506167,
+ "loss": 2.107178211212158,
+ "step": 1320
+ },
+ {
+ "epoch": 18.34843205574913,
+ "grad_norm": 0.0534445196390152,
+ "learning_rate": 0.0004390214684046111,
+ "loss": 2.1187639236450195,
+ "step": 1321
+ },
+ {
+ "epoch": 18.362369337979093,
+ "grad_norm": 0.056623175740242004,
+ "learning_rate": 0.00043878442882412015,
+ "loss": 2.1187937259674072,
+ "step": 1322
+ },
+ {
+ "epoch": 18.37630662020906,
+ "grad_norm": 0.05681455507874489,
+ "learning_rate": 0.00043854727984456746,
+ "loss": 2.1196155548095703,
+ "step": 1323
+ },
+ {
+ "epoch": 18.390243902439025,
+ "grad_norm": 0.05910615995526314,
+ "learning_rate": 0.0004383100216570197,
+ "loss": 2.1284592151641846,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40418118466899,
+ "grad_norm": 0.05617159232497215,
+ "learning_rate": 0.0004380726544526316,
+ "loss": 2.1014914512634277,
+ "step": 1325
+ },
+ {
+ "epoch": 18.418118466898953,
+ "grad_norm": 0.05503782257437706,
+ "learning_rate": 0.00043783517842264534,
+ "loss": 2.113588333129883,
+ "step": 1326
+ },
+ {
+ "epoch": 18.43205574912892,
+ "grad_norm": 0.0550033375620842,
+ "learning_rate": 0.0004375975937583914,
+ "loss": 2.137667655944824,
+ "step": 1327
+ },
+ {
+ "epoch": 18.445993031358885,
+ "grad_norm": 0.05703869089484215,
+ "learning_rate": 0.0004373599006512871,
+ "loss": 2.1277079582214355,
+ "step": 1328
+ },
+ {
+ "epoch": 18.45993031358885,
+ "grad_norm": 0.05607055500149727,
+ "learning_rate": 0.00043712209929283786,
+ "loss": 2.118278980255127,
+ "step": 1329
+ },
+ {
+ "epoch": 18.473867595818817,
+ "grad_norm": 0.054871171712875366,
+ "learning_rate": 0.00043688418987463567,
+ "loss": 2.123734951019287,
+ "step": 1330
+ },
+ {
+ "epoch": 18.48780487804878,
+ "grad_norm": 0.053621839731931686,
+ "learning_rate": 0.00043664617258835995,
+ "loss": 2.146944761276245,
+ "step": 1331
+ },
+ {
+ "epoch": 18.501742160278745,
+ "grad_norm": 0.057774804532527924,
+ "learning_rate": 0.0004364080476257769,
+ "loss": 2.131260871887207,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51567944250871,
+ "grad_norm": 0.06485384702682495,
+ "learning_rate": 0.0004361698151787396,
+ "loss": 2.118954658508301,
+ "step": 1333
+ },
+ {
+ "epoch": 18.529616724738677,
+ "grad_norm": 0.06402954459190369,
+ "learning_rate": 0.0004359314754391874,
+ "loss": 2.1208548545837402,
+ "step": 1334
+ },
+ {
+ "epoch": 18.54355400696864,
+ "grad_norm": 0.05448993667960167,
+ "learning_rate": 0.0004356930285991466,
+ "loss": 2.129748821258545,
+ "step": 1335
+ },
+ {
+ "epoch": 18.557491289198605,
+ "grad_norm": 0.05529506132006645,
+ "learning_rate": 0.00043545447485072926,
+ "loss": 2.130112648010254,
+ "step": 1336
+ },
+ {
+ "epoch": 18.571428571428573,
+ "grad_norm": 0.054673660546541214,
+ "learning_rate": 0.000435215814386134,
+ "loss": 2.1194992065429688,
+ "step": 1337
+ },
+ {
+ "epoch": 18.585365853658537,
+ "grad_norm": 0.053571924567222595,
+ "learning_rate": 0.0004349770473976453,
+ "loss": 2.128544569015503,
+ "step": 1338
+ },
+ {
+ "epoch": 18.5993031358885,
+ "grad_norm": 0.05538960546255112,
+ "learning_rate": 0.0004347381740776332,
+ "loss": 2.1390209197998047,
+ "step": 1339
+ },
+ {
+ "epoch": 18.613240418118465,
+ "grad_norm": 0.058218978345394135,
+ "learning_rate": 0.0004344991946185538,
+ "loss": 2.134378433227539,
+ "step": 1340
+ },
+ {
+ "epoch": 18.627177700348433,
+ "grad_norm": 0.058131180703639984,
+ "learning_rate": 0.00043426010921294837,
+ "loss": 2.142953872680664,
+ "step": 1341
+ },
+ {
+ "epoch": 18.641114982578397,
+ "grad_norm": 0.05974142625927925,
+ "learning_rate": 0.0004340209180534438,
+ "loss": 2.1428234577178955,
+ "step": 1342
+ },
+ {
+ "epoch": 18.65505226480836,
+ "grad_norm": 0.057082053273916245,
+ "learning_rate": 0.000433781621332752,
+ "loss": 2.1265363693237305,
+ "step": 1343
+ },
+ {
+ "epoch": 18.66898954703833,
+ "grad_norm": 0.056070681661367416,
+ "learning_rate": 0.0004335422192436703,
+ "loss": 2.1269443035125732,
+ "step": 1344
+ },
+ {
+ "epoch": 18.682926829268293,
+ "grad_norm": 0.05955109745264053,
+ "learning_rate": 0.00043330271197908023,
+ "loss": 2.135490655899048,
+ "step": 1345
+ },
+ {
+ "epoch": 18.696864111498257,
+ "grad_norm": 0.05841276794672012,
+ "learning_rate": 0.00043306309973194874,
+ "loss": 2.124037027359009,
+ "step": 1346
+ },
+ {
+ "epoch": 18.71080139372822,
+ "grad_norm": 0.05835938826203346,
+ "learning_rate": 0.000432823382695327,
+ "loss": 2.145099401473999,
+ "step": 1347
+ },
+ {
+ "epoch": 18.72473867595819,
+ "grad_norm": 0.05872299149632454,
+ "learning_rate": 0.0004325835610623508,
+ "loss": 2.142068862915039,
+ "step": 1348
+ },
+ {
+ "epoch": 18.738675958188153,
+ "grad_norm": 0.05572303384542465,
+ "learning_rate": 0.00043234363502623994,
+ "loss": 2.1334776878356934,
+ "step": 1349
+ },
+ {
+ "epoch": 18.752613240418118,
+ "grad_norm": 0.054891571402549744,
+ "learning_rate": 0.0004321036047802985,
+ "loss": 2.1316189765930176,
+ "step": 1350
+ },
+ {
+ "epoch": 18.766550522648085,
+ "grad_norm": 0.052549563348293304,
+ "learning_rate": 0.00043186347051791466,
+ "loss": 2.1244778633117676,
+ "step": 1351
+ },
+ {
+ "epoch": 18.78048780487805,
+ "grad_norm": 0.05204412341117859,
+ "learning_rate": 0.0004316232324325602,
+ "loss": 2.142746925354004,
+ "step": 1352
+ },
+ {
+ "epoch": 18.794425087108014,
+ "grad_norm": 0.05793134495615959,
+ "learning_rate": 0.0004313828907177906,
+ "loss": 2.1276097297668457,
+ "step": 1353
+ },
+ {
+ "epoch": 18.808362369337978,
+ "grad_norm": 0.057822611182928085,
+ "learning_rate": 0.0004311424455672448,
+ "loss": 2.1336703300476074,
+ "step": 1354
+ },
+ {
+ "epoch": 18.822299651567945,
+ "grad_norm": 0.05854838341474533,
+ "learning_rate": 0.0004309018971746451,
+ "loss": 2.1333818435668945,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83623693379791,
+ "grad_norm": 0.062328968197107315,
+ "learning_rate": 0.000430661245733797,
+ "loss": 2.1374850273132324,
+ "step": 1356
+ },
+ {
+ "epoch": 18.850174216027874,
+ "grad_norm": 0.06169719994068146,
+ "learning_rate": 0.00043042049143858914,
+ "loss": 2.1548595428466797,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86411149825784,
+ "grad_norm": 0.05472258850932121,
+ "learning_rate": 0.0004301796344829928,
+ "loss": 2.146165132522583,
+ "step": 1358
+ },
+ {
+ "epoch": 18.878048780487806,
+ "grad_norm": 0.054294172674417496,
+ "learning_rate": 0.00042993867506106205,
+ "loss": 2.1387805938720703,
+ "step": 1359
+ },
+ {
+ "epoch": 18.89198606271777,
+ "grad_norm": 0.05388478934764862,
+ "learning_rate": 0.0004296976133669336,
+ "loss": 2.1439051628112793,
+ "step": 1360
+ },
+ {
+ "epoch": 18.905923344947734,
+ "grad_norm": 0.05328727513551712,
+ "learning_rate": 0.0004294564495948266,
+ "loss": 2.1475648880004883,
+ "step": 1361
+ },
+ {
+ "epoch": 18.9198606271777,
+ "grad_norm": 0.05454222857952118,
+ "learning_rate": 0.0004292151839390421,
+ "loss": 2.1299092769622803,
+ "step": 1362
+ },
+ {
+ "epoch": 18.933797909407666,
+ "grad_norm": 0.0570770688354969,
+ "learning_rate": 0.0004289738165939636,
+ "loss": 2.119565486907959,
+ "step": 1363
+ },
+ {
+ "epoch": 18.94773519163763,
+ "grad_norm": 0.05483704432845116,
+ "learning_rate": 0.00042873234775405654,
+ "loss": 2.1410772800445557,
+ "step": 1364
+ },
+ {
+ "epoch": 18.961672473867594,
+ "grad_norm": 0.05490054935216904,
+ "learning_rate": 0.00042849077761386784,
+ "loss": 2.144008159637451,
+ "step": 1365
+ },
+ {
+ "epoch": 18.975609756097562,
+ "grad_norm": 0.04970669746398926,
+ "learning_rate": 0.0004282491063680263,
+ "loss": 2.136209487915039,
+ "step": 1366
+ },
+ {
+ "epoch": 18.989547038327526,
+ "grad_norm": 0.05208569020032883,
+ "learning_rate": 0.00042800733421124207,
+ "loss": 2.147538185119629,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.04167763888835907,
+ "learning_rate": 0.00042776546133830646,
+ "loss": 1.5960297584533691,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 0.6202362775802612,
+ "eval_runtime": 40.6052,
+ "eval_samples_per_second": 60.14,
+ "eval_steps_per_second": 0.493,
+ "step": 1368
+ },
+ {
+ "epoch": 19.013937282229964,
+ "grad_norm": 0.052768524736166,
+ "learning_rate": 0.00042752348794409227,
+ "loss": 2.0851144790649414,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027874564459932,
+ "grad_norm": 0.058598630130290985,
+ "learning_rate": 0.00042728141422355287,
+ "loss": 2.088078737258911,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041811846689896,
+ "grad_norm": 0.05658827722072601,
+ "learning_rate": 0.0004270392403717229,
+ "loss": 2.0905416011810303,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05574912891986,
+ "grad_norm": 0.060161419212818146,
+ "learning_rate": 0.00042679696658371743,
+ "loss": 2.0762481689453125,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069686411149824,
+ "grad_norm": 0.05761908367276192,
+ "learning_rate": 0.00042655459305473195,
+ "loss": 2.0815508365631104,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083623693379792,
+ "grad_norm": 0.05942010134458542,
+ "learning_rate": 0.0004263121199800425,
+ "loss": 2.081665515899658,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097560975609756,
+ "grad_norm": 0.06076835095882416,
+ "learning_rate": 0.0004260695475550054,
+ "loss": 2.089571952819824,
+ "step": 1375
+ },
+ {
+ "epoch": 19.11149825783972,
+ "grad_norm": 0.05942385643720627,
+ "learning_rate": 0.0004258268759750566,
+ "loss": 2.0848846435546875,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125435540069688,
+ "grad_norm": 0.06429574638605118,
+ "learning_rate": 0.0004255841054357124,
+ "loss": 2.0932364463806152,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139372822299652,
+ "grad_norm": 0.06300369650125504,
+ "learning_rate": 0.0004253412361325686,
+ "loss": 2.087846517562866,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153310104529616,
+ "grad_norm": 0.06136588379740715,
+ "learning_rate": 0.0004250982682613006,
+ "loss": 2.0813498497009277,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16724738675958,
+ "grad_norm": 0.05989256873726845,
+ "learning_rate": 0.00042485520201766315,
+ "loss": 2.075523614883423,
+ "step": 1380
+ },
+ {
+ "epoch": 19.181184668989548,
+ "grad_norm": 0.058910951018333435,
+ "learning_rate": 0.0004246120375974905,
+ "loss": 2.0861804485321045,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195121951219512,
+ "grad_norm": 0.06127007305622101,
+ "learning_rate": 0.0004243687751966957,
+ "loss": 2.078918218612671,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209059233449477,
+ "grad_norm": 0.0580737479031086,
+ "learning_rate": 0.0004241254150112708,
+ "loss": 2.0984582901000977,
+ "step": 1383
+ },
+ {
+ "epoch": 19.222996515679444,
+ "grad_norm": 0.055373016744852066,
+ "learning_rate": 0.000423881957237287,
+ "loss": 2.0958027839660645,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23693379790941,
+ "grad_norm": 0.05787358433008194,
+ "learning_rate": 0.0004236384020708935,
+ "loss": 2.0821218490600586,
+ "step": 1385
+ },
+ {
+ "epoch": 19.250871080139373,
+ "grad_norm": 0.06461151689291,
+ "learning_rate": 0.0004233947497083185,
+ "loss": 2.095828056335449,
+ "step": 1386
+ },
+ {
+ "epoch": 19.264808362369337,
+ "grad_norm": 0.061279844492673874,
+ "learning_rate": 0.0004231510003458682,
+ "loss": 2.091866970062256,
+ "step": 1387
+ },
+ {
+ "epoch": 19.278745644599304,
+ "grad_norm": 0.05671652406454086,
+ "learning_rate": 0.0004229071541799272,
+ "loss": 2.1064250469207764,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29268292682927,
+ "grad_norm": 0.06336323916912079,
+ "learning_rate": 0.0004226632114069578,
+ "loss": 2.1064412593841553,
+ "step": 1389
+ },
+ {
+ "epoch": 19.306620209059233,
+ "grad_norm": 0.06304406374692917,
+ "learning_rate": 0.0004224191722235004,
+ "loss": 2.107055187225342,
+ "step": 1390
+ },
+ {
+ "epoch": 19.320557491289197,
+ "grad_norm": 0.06382641196250916,
+ "learning_rate": 0.00042217503682617286,
+ "loss": 2.0970683097839355,
+ "step": 1391
+ },
+ {
+ "epoch": 19.334494773519165,
+ "grad_norm": 0.06603065878152847,
+ "learning_rate": 0.0004219308054116706,
+ "loss": 2.101444721221924,
+ "step": 1392
+ },
+ {
+ "epoch": 19.34843205574913,
+ "grad_norm": 0.06299275159835815,
+ "learning_rate": 0.0004216864781767667,
+ "loss": 2.091536521911621,
+ "step": 1393
+ },
+ {
+ "epoch": 19.362369337979093,
+ "grad_norm": 0.057733237743377686,
+ "learning_rate": 0.00042144205531831103,
+ "loss": 2.1091341972351074,
+ "step": 1394
+ },
+ {
+ "epoch": 19.37630662020906,
+ "grad_norm": 0.05802382901310921,
+ "learning_rate": 0.0004211975370332308,
+ "loss": 2.0830318927764893,
+ "step": 1395
+ },
+ {
+ "epoch": 19.390243902439025,
+ "grad_norm": 0.0592074953019619,
+ "learning_rate": 0.00042095292351852976,
+ "loss": 2.0685744285583496,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40418118466899,
+ "grad_norm": 0.056426629424095154,
+ "learning_rate": 0.0004207082149712888,
+ "loss": 2.0968804359436035,
+ "step": 1397
+ },
+ {
+ "epoch": 19.418118466898953,
+ "grad_norm": 0.05799269303679466,
+ "learning_rate": 0.0004204634115886651,
+ "loss": 2.092805862426758,
+ "step": 1398
+ },
+ {
+ "epoch": 19.43205574912892,
+ "grad_norm": 0.05895870178937912,
+ "learning_rate": 0.0004202185135678924,
+ "loss": 2.108272075653076,
+ "step": 1399
+ },
+ {
+ "epoch": 19.445993031358885,
+ "grad_norm": 0.05725407600402832,
+ "learning_rate": 0.00041997352110628033,
+ "loss": 2.100399971008301,
+ "step": 1400
+ },
+ {
+ "epoch": 19.45993031358885,
+ "grad_norm": 0.05726857855916023,
+ "learning_rate": 0.00041972843440121527,
+ "loss": 2.1097989082336426,
+ "step": 1401
+ },
+ {
+ "epoch": 19.473867595818817,
+ "grad_norm": 0.05829065665602684,
+ "learning_rate": 0.00041948325365015894,
+ "loss": 2.1080002784729004,
+ "step": 1402
+ },
+ {
+ "epoch": 19.48780487804878,
+ "grad_norm": 0.057527437806129456,
+ "learning_rate": 0.0004192379790506491,
+ "loss": 2.1043288707733154,
+ "step": 1403
+ },
+ {
+ "epoch": 19.501742160278745,
+ "grad_norm": 0.06426865607500076,
+ "learning_rate": 0.00041899261080029904,
+ "loss": 2.117246627807617,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51567944250871,
+ "grad_norm": 0.0638858824968338,
+ "learning_rate": 0.00041874714909679754,
+ "loss": 2.1121253967285156,
+ "step": 1405
+ },
+ {
+ "epoch": 19.529616724738677,
+ "grad_norm": 0.05847329646348953,
+ "learning_rate": 0.00041850159413790863,
+ "loss": 2.121541976928711,
+ "step": 1406
+ },
+ {
+ "epoch": 19.54355400696864,
+ "grad_norm": 0.056684136390686035,
+ "learning_rate": 0.0004182559461214715,
+ "loss": 2.1145877838134766,
+ "step": 1407
+ },
+ {
+ "epoch": 19.557491289198605,
+ "grad_norm": 0.06320773810148239,
+ "learning_rate": 0.0004180102052454004,
+ "loss": 2.1046271324157715,
+ "step": 1408
+ },
+ {
+ "epoch": 19.571428571428573,
+ "grad_norm": 0.06170722842216492,
+ "learning_rate": 0.00041776437170768427,
+ "loss": 2.1137208938598633,
+ "step": 1409
+ },
+ {
+ "epoch": 19.585365853658537,
+ "grad_norm": 0.05700785666704178,
+ "learning_rate": 0.0004175184457063868,
+ "loss": 2.10615611076355,
+ "step": 1410
+ },
+ {
+ "epoch": 19.5993031358885,
+ "grad_norm": 0.05721168965101242,
+ "learning_rate": 0.000417272427439646,
+ "loss": 2.1167562007904053,
+ "step": 1411
+ },
+ {
+ "epoch": 19.613240418118465,
+ "grad_norm": 0.0627029612660408,
+ "learning_rate": 0.00041702631710567443,
+ "loss": 2.1217145919799805,
+ "step": 1412
+ },
+ {
+ "epoch": 19.627177700348433,
+ "grad_norm": 0.061327043920755386,
+ "learning_rate": 0.00041678011490275875,
+ "loss": 2.1209311485290527,
+ "step": 1413
+ },
+ {
+ "epoch": 19.641114982578397,
+ "grad_norm": 0.06154641881585121,
+ "learning_rate": 0.00041653382102925957,
+ "loss": 2.112914562225342,
+ "step": 1414
+ },
+ {
+ "epoch": 19.65505226480836,
+ "grad_norm": 0.057539358735084534,
+ "learning_rate": 0.00041628743568361147,
+ "loss": 2.120575428009033,
+ "step": 1415
+ },
+ {
+ "epoch": 19.66898954703833,
+ "grad_norm": 0.057371243834495544,
+ "learning_rate": 0.00041604095906432265,
+ "loss": 2.117800235748291,
+ "step": 1416
+ },
+ {
+ "epoch": 19.682926829268293,
+ "grad_norm": 0.059675198048353195,
+ "learning_rate": 0.00041579439136997493,
+ "loss": 2.1061432361602783,
+ "step": 1417
+ },
+ {
+ "epoch": 19.696864111498257,
+ "grad_norm": 0.057770736515522,
+ "learning_rate": 0.00041554773279922337,
+ "loss": 2.1168129444122314,
+ "step": 1418
+ },
+ {
+ "epoch": 19.71080139372822,
+ "grad_norm": 0.054123856127262115,
+ "learning_rate": 0.0004153009835507963,
+ "loss": 2.118847370147705,
+ "step": 1419
+ },
+ {
+ "epoch": 19.72473867595819,
+ "grad_norm": 0.05700135976076126,
+ "learning_rate": 0.0004150541438234952,
+ "loss": 2.0995051860809326,
+ "step": 1420
+ },
+ {
+ "epoch": 19.738675958188153,
+ "grad_norm": 0.05754461511969566,
+ "learning_rate": 0.00041480721381619435,
+ "loss": 2.1250476837158203,
+ "step": 1421
+ },
+ {
+ "epoch": 19.752613240418118,
+ "grad_norm": 0.055894333869218826,
+ "learning_rate": 0.00041456019372784086,
+ "loss": 2.115142345428467,
+ "step": 1422
+ },
+ {
+ "epoch": 19.766550522648085,
+ "grad_norm": 0.05652809888124466,
+ "learning_rate": 0.0004143130837574543,
+ "loss": 2.1061787605285645,
+ "step": 1423
+ },
+ {
+ "epoch": 19.78048780487805,
+ "grad_norm": 0.058509763330221176,
+ "learning_rate": 0.0004140658841041267,
+ "loss": 2.1187353134155273,
+ "step": 1424
+ },
+ {
+ "epoch": 19.794425087108014,
+ "grad_norm": 0.06004858762025833,
+ "learning_rate": 0.0004138185949670224,
+ "loss": 2.131927251815796,
+ "step": 1425
+ },
+ {
+ "epoch": 19.808362369337978,
+ "grad_norm": 0.05746140703558922,
+ "learning_rate": 0.00041357121654537766,
+ "loss": 2.114919662475586,
+ "step": 1426
+ },
+ {
+ "epoch": 19.822299651567945,
+ "grad_norm": 0.05890800058841705,
+ "learning_rate": 0.00041332374903850104,
+ "loss": 2.109203815460205,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83623693379791,
+ "grad_norm": 0.05798939988017082,
+ "learning_rate": 0.00041307619264577234,
+ "loss": 2.1246418952941895,
+ "step": 1428
+ },
+ {
+ "epoch": 19.850174216027874,
+ "grad_norm": 0.05782999098300934,
+ "learning_rate": 0.0004128285475666436,
+ "loss": 2.1183271408081055,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86411149825784,
+ "grad_norm": 0.05817651003599167,
+ "learning_rate": 0.0004125808140006378,
+ "loss": 2.1114745140075684,
+ "step": 1430
+ },
+ {
+ "epoch": 19.878048780487806,
+ "grad_norm": 0.058334168046712875,
+ "learning_rate": 0.0004123329921473494,
+ "loss": 2.121798515319824,
+ "step": 1431
+ },
+ {
+ "epoch": 19.89198606271777,
+ "grad_norm": 0.05437908321619034,
+ "learning_rate": 0.0004120850822064439,
+ "loss": 2.107545852661133,
+ "step": 1432
+ },
+ {
+ "epoch": 19.905923344947734,
+ "grad_norm": 0.05461714416742325,
+ "learning_rate": 0.00041183708437765815,
+ "loss": 2.126991033554077,
+ "step": 1433
+ },
+ {
+ "epoch": 19.9198606271777,
+ "grad_norm": 0.058922551572322845,
+ "learning_rate": 0.00041158899886079925,
+ "loss": 2.1250553131103516,
+ "step": 1434
+ },
+ {
+ "epoch": 19.933797909407666,
+ "grad_norm": 0.05759972706437111,
+ "learning_rate": 0.00041134082585574534,
+ "loss": 2.1215944290161133,
+ "step": 1435
+ },
+ {
+ "epoch": 19.94773519163763,
+ "grad_norm": 0.057332854717969894,
+ "learning_rate": 0.00041109256556244497,
+ "loss": 2.1251895427703857,
+ "step": 1436
+ },
+ {
+ "epoch": 19.961672473867594,
+ "grad_norm": 0.056731462478637695,
+ "learning_rate": 0.00041084421818091685,
+ "loss": 2.109805107116699,
+ "step": 1437
+ },
+ {
+ "epoch": 19.975609756097562,
+ "grad_norm": 0.05827569589018822,
+ "learning_rate": 0.0004105957839112501,
+ "loss": 2.1196374893188477,
+ "step": 1438
+ },
+ {
+ "epoch": 19.989547038327526,
+ "grad_norm": 0.05641156807541847,
+ "learning_rate": 0.0004103472629536036,
+ "loss": 2.1140520572662354,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.04525474086403847,
+ "learning_rate": 0.0004100986555082063,
+ "loss": 1.585239052772522,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 0.621873140335083,
+ "eval_runtime": 40.0081,
+ "eval_samples_per_second": 61.038,
+ "eval_steps_per_second": 0.5,
+ "step": 1440
+ },
+ {
+ "epoch": 20.013937282229964,
+ "grad_norm": 0.05569886416196823,
+ "learning_rate": 0.0004098499617753567,
+ "loss": 2.0722551345825195,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027874564459932,
+ "grad_norm": 0.06597832590341568,
+ "learning_rate": 0.00040960118195542273,
+ "loss": 2.0547425746917725,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041811846689896,
+ "grad_norm": 0.058331988751888275,
+ "learning_rate": 0.00040935231624884204,
+ "loss": 2.0709681510925293,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05574912891986,
+ "grad_norm": 0.05631960928440094,
+ "learning_rate": 0.0004091033648561211,
+ "loss": 2.072939395904541,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069686411149824,
+ "grad_norm": 0.06110234558582306,
+ "learning_rate": 0.00040885432797783543,
+ "loss": 2.0694265365600586,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083623693379792,
+ "grad_norm": 0.05794587731361389,
+ "learning_rate": 0.0004086052058146297,
+ "loss": 2.0740833282470703,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097560975609756,
+ "grad_norm": 0.05939078703522682,
+ "learning_rate": 0.00040835599856721725,
+ "loss": 2.0684261322021484,
+ "step": 1447
+ },
+ {
+ "epoch": 20.11149825783972,
+ "grad_norm": 0.05748258903622627,
+ "learning_rate": 0.00040810670643637954,
+ "loss": 2.0731234550476074,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125435540069688,
+ "grad_norm": 0.05458281934261322,
+ "learning_rate": 0.000407857329622967,
+ "loss": 2.059173583984375,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139372822299652,
+ "grad_norm": 0.0548420324921608,
+ "learning_rate": 0.00040760786832789786,
+ "loss": 2.067276954650879,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153310104529616,
+ "grad_norm": 0.05886463448405266,
+ "learning_rate": 0.0004073583227521588,
+ "loss": 2.0733237266540527,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16724738675958,
+ "grad_norm": 0.056409262120723724,
+ "learning_rate": 0.0004071086930968039,
+ "loss": 2.0705678462982178,
+ "step": 1452
+ },
+ {
+ "epoch": 20.181184668989548,
+ "grad_norm": 0.06016712635755539,
+ "learning_rate": 0.00040685897956295545,
+ "loss": 2.0607805252075195,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195121951219512,
+ "grad_norm": 0.06224392354488373,
+ "learning_rate": 0.0004066091823518031,
+ "loss": 2.0710277557373047,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209059233449477,
+ "grad_norm": 0.055681005120277405,
+ "learning_rate": 0.00040635930166460385,
+ "loss": 2.081244468688965,
+ "step": 1455
+ },
+ {
+ "epoch": 20.222996515679444,
+ "grad_norm": 0.059053998440504074,
+ "learning_rate": 0.00040610933770268226,
+ "loss": 2.0709948539733887,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23693379790941,
+ "grad_norm": 0.06237979233264923,
+ "learning_rate": 0.00040585929066742964,
+ "loss": 2.0689477920532227,
+ "step": 1457
+ },
+ {
+ "epoch": 20.250871080139373,
+ "grad_norm": 0.06198735907673836,
+ "learning_rate": 0.00040560916076030435,
+ "loss": 2.0899670124053955,
+ "step": 1458
+ },
+ {
+ "epoch": 20.264808362369337,
+ "grad_norm": 0.06024961546063423,
+ "learning_rate": 0.00040535894818283154,
+ "loss": 2.0820512771606445,
+ "step": 1459
+ },
+ {
+ "epoch": 20.278745644599304,
+ "grad_norm": 0.06023218855261803,
+ "learning_rate": 0.0004051086531366031,
+ "loss": 2.062567710876465,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29268292682927,
+ "grad_norm": 0.06317745894193649,
+ "learning_rate": 0.000404858275823277,
+ "loss": 2.0713205337524414,
+ "step": 1461
+ },
+ {
+ "epoch": 20.306620209059233,
+ "grad_norm": 0.06885461509227753,
+ "learning_rate": 0.000404607816444578,
+ "loss": 2.0726189613342285,
+ "step": 1462
+ },
+ {
+ "epoch": 20.320557491289197,
+ "grad_norm": 0.06577035039663315,
+ "learning_rate": 0.00040435727520229644,
+ "loss": 2.0791847705841064,
+ "step": 1463
+ },
+ {
+ "epoch": 20.334494773519165,
+ "grad_norm": 0.062136828899383545,
+ "learning_rate": 0.0004041066522982891,
+ "loss": 2.072225570678711,
+ "step": 1464
+ },
+ {
+ "epoch": 20.34843205574913,
+ "grad_norm": 0.06487686187028885,
+ "learning_rate": 0.000403855947934478,
+ "loss": 2.0821399688720703,
+ "step": 1465
+ },
+ {
+ "epoch": 20.362369337979093,
+ "grad_norm": 0.06462955474853516,
+ "learning_rate": 0.00040360516231285155,
+ "loss": 2.0996780395507812,
+ "step": 1466
+ },
+ {
+ "epoch": 20.37630662020906,
+ "grad_norm": 0.06359101086854935,
+ "learning_rate": 0.00040335429563546286,
+ "loss": 2.083282470703125,
+ "step": 1467
+ },
+ {
+ "epoch": 20.390243902439025,
+ "grad_norm": 0.05839303508400917,
+ "learning_rate": 0.0004031033481044308,
+ "loss": 2.0780766010284424,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40418118466899,
+ "grad_norm": 0.06317011266946793,
+ "learning_rate": 0.00040285231992193924,
+ "loss": 2.0861825942993164,
+ "step": 1469
+ },
+ {
+ "epoch": 20.418118466898953,
+ "grad_norm": 0.06130538880825043,
+ "learning_rate": 0.0004026012112902372,
+ "loss": 2.0851950645446777,
+ "step": 1470
+ },
+ {
+ "epoch": 20.43205574912892,
+ "grad_norm": 0.06157645955681801,
+ "learning_rate": 0.0004023500224116381,
+ "loss": 2.0819547176361084,
+ "step": 1471
+ },
+ {
+ "epoch": 20.445993031358885,
+ "grad_norm": 0.059893663972616196,
+ "learning_rate": 0.00040209875348852037,
+ "loss": 2.080601215362549,
+ "step": 1472
+ },
+ {
+ "epoch": 20.45993031358885,
+ "grad_norm": 0.05882779136300087,
+ "learning_rate": 0.00040184740472332705,
+ "loss": 2.0791540145874023,
+ "step": 1473
+ },
+ {
+ "epoch": 20.473867595818817,
+ "grad_norm": 0.058982688933610916,
+ "learning_rate": 0.000401595976318565,
+ "loss": 2.0775556564331055,
+ "step": 1474
+ },
+ {
+ "epoch": 20.48780487804878,
+ "grad_norm": 0.060487501323223114,
+ "learning_rate": 0.00040134446847680584,
+ "loss": 2.0972094535827637,
+ "step": 1475
+ },
+ {
+ "epoch": 20.501742160278745,
+ "grad_norm": 0.06333193182945251,
+ "learning_rate": 0.0004010928814006846,
+ "loss": 2.084468364715576,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51567944250871,
+ "grad_norm": 0.0660906732082367,
+ "learning_rate": 0.0004008412152929007,
+ "loss": 2.0966973304748535,
+ "step": 1477
+ },
+ {
+ "epoch": 20.529616724738677,
+ "grad_norm": 0.06255893409252167,
+ "learning_rate": 0.0004005894703562168,
+ "loss": 2.078449249267578,
+ "step": 1478
+ },
+ {
+ "epoch": 20.54355400696864,
+ "grad_norm": 0.05733060464262962,
+ "learning_rate": 0.0004003376467934593,
+ "loss": 2.0744051933288574,
+ "step": 1479
+ },
+ {
+ "epoch": 20.557491289198605,
+ "grad_norm": 0.060031335800886154,
+ "learning_rate": 0.0004000857448075178,
+ "loss": 2.093151569366455,
+ "step": 1480
+ },
+ {
+ "epoch": 20.571428571428573,
+ "grad_norm": 0.06155857443809509,
+ "learning_rate": 0.0003998337646013455,
+ "loss": 2.0912318229675293,
+ "step": 1481
+ },
+ {
+ "epoch": 20.585365853658537,
+ "grad_norm": 0.06131042540073395,
+ "learning_rate": 0.00039958170637795806,
+ "loss": 2.091536045074463,
+ "step": 1482
+ },
+ {
+ "epoch": 20.5993031358885,
+ "grad_norm": 0.056901413947343826,
+ "learning_rate": 0.0003993295703404341,
+ "loss": 2.087958335876465,
+ "step": 1483
+ },
+ {
+ "epoch": 20.613240418118465,
+ "grad_norm": 0.05618821084499359,
+ "learning_rate": 0.00039907735669191544,
+ "loss": 2.0817902088165283,
+ "step": 1484
+ },
+ {
+ "epoch": 20.627177700348433,
+ "grad_norm": 0.0622171014547348,
+ "learning_rate": 0.00039882506563560573,
+ "loss": 2.102149486541748,
+ "step": 1485
+ },
+ {
+ "epoch": 20.641114982578397,
+ "grad_norm": 0.05873639136552811,
+ "learning_rate": 0.0003985726973747715,
+ "loss": 2.0869569778442383,
+ "step": 1486
+ },
+ {
+ "epoch": 20.65505226480836,
+ "grad_norm": 0.05968155711889267,
+ "learning_rate": 0.00039832025211274113,
+ "loss": 2.0925517082214355,
+ "step": 1487
+ },
+ {
+ "epoch": 20.66898954703833,
+ "grad_norm": 0.05784456059336662,
+ "learning_rate": 0.00039806773005290544,
+ "loss": 2.092297077178955,
+ "step": 1488
+ },
+ {
+ "epoch": 20.682926829268293,
+ "grad_norm": 0.05630253255367279,
+ "learning_rate": 0.0003978151313987168,
+ "loss": 2.0782198905944824,
+ "step": 1489
+ },
+ {
+ "epoch": 20.696864111498257,
+ "grad_norm": 0.05917421728372574,
+ "learning_rate": 0.00039756245635368933,
+ "loss": 2.0907468795776367,
+ "step": 1490
+ },
+ {
+ "epoch": 20.71080139372822,
+ "grad_norm": 0.05753430724143982,
+ "learning_rate": 0.00039730970512139876,
+ "loss": 2.0973947048187256,
+ "step": 1491
+ },
+ {
+ "epoch": 20.72473867595819,
+ "grad_norm": 0.05667796730995178,
+ "learning_rate": 0.0003970568779054822,
+ "loss": 2.0887746810913086,
+ "step": 1492
+ },
+ {
+ "epoch": 20.738675958188153,
+ "grad_norm": 0.054731108248233795,
+ "learning_rate": 0.000396803974909638,
+ "loss": 2.103682518005371,
+ "step": 1493
+ },
+ {
+ "epoch": 20.752613240418118,
+ "grad_norm": 0.05500303581357002,
+ "learning_rate": 0.0003965509963376255,
+ "loss": 2.075216293334961,
+ "step": 1494
+ },
+ {
+ "epoch": 20.766550522648085,
+ "grad_norm": 0.05450119450688362,
+ "learning_rate": 0.000396297942393265,
+ "loss": 2.0977306365966797,
+ "step": 1495
+ },
+ {
+ "epoch": 20.78048780487805,
+ "grad_norm": 0.05592454969882965,
+ "learning_rate": 0.0003960448132804375,
+ "loss": 2.0991897583007812,
+ "step": 1496
+ },
+ {
+ "epoch": 20.794425087108014,
+ "grad_norm": 0.05683634430170059,
+ "learning_rate": 0.0003957916092030845,
+ "loss": 2.0911078453063965,
+ "step": 1497
+ },
+ {
+ "epoch": 20.808362369337978,
+ "grad_norm": 0.058022066950798035,
+ "learning_rate": 0.00039553833036520803,
+ "loss": 2.0965914726257324,
+ "step": 1498
+ },
+ {
+ "epoch": 20.822299651567945,
+ "grad_norm": 0.056369319558143616,
+ "learning_rate": 0.0003952849769708702,
+ "loss": 2.0945000648498535,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83623693379791,
+ "grad_norm": 0.05577847734093666,
+ "learning_rate": 0.0003950315492241932,
+ "loss": 2.1021676063537598,
+ "step": 1500
+ },
+ {
+ "epoch": 20.850174216027874,
+ "grad_norm": 0.055830225348472595,
+ "learning_rate": 0.0003947780473293593,
+ "loss": 2.121771812438965,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86411149825784,
+ "grad_norm": 0.05487843230366707,
+ "learning_rate": 0.00039452447149061054,
+ "loss": 2.0852222442626953,
+ "step": 1502
+ },
+ {
+ "epoch": 20.878048780487806,
+ "grad_norm": 0.05502147600054741,
+ "learning_rate": 0.0003942708219122481,
+ "loss": 2.1024603843688965,
+ "step": 1503
+ },
+ {
+ "epoch": 20.89198606271777,
+ "grad_norm": 0.05633259564638138,
+ "learning_rate": 0.000394017098798633,
+ "loss": 2.0984888076782227,
+ "step": 1504
+ },
+ {
+ "epoch": 20.905923344947734,
+ "grad_norm": 0.05579936131834984,
+ "learning_rate": 0.0003937633023541854,
+ "loss": 2.1036808490753174,
+ "step": 1505
+ },
+ {
+ "epoch": 20.9198606271777,
+ "grad_norm": 0.05743060261011124,
+ "learning_rate": 0.0003935094327833845,
+ "loss": 2.079979658126831,
+ "step": 1506
+ },
+ {
+ "epoch": 20.933797909407666,
+ "grad_norm": 0.054715096950531006,
+ "learning_rate": 0.0003932554902907683,
+ "loss": 2.086301803588867,
+ "step": 1507
+ },
+ {
+ "epoch": 20.94773519163763,
+ "grad_norm": 0.05395551398396492,
+ "learning_rate": 0.0003930014750809338,
+ "loss": 2.1051979064941406,
+ "step": 1508
+ },
+ {
+ "epoch": 20.961672473867594,
+ "grad_norm": 0.05604757368564606,
+ "learning_rate": 0.0003927473873585365,
+ "loss": 2.1195569038391113,
+ "step": 1509
+ },
+ {
+ "epoch": 20.975609756097562,
+ "grad_norm": 0.06016363948583603,
+ "learning_rate": 0.0003924932273282903,
+ "loss": 2.0948679447174072,
+ "step": 1510
+ },
+ {
+ "epoch": 20.989547038327526,
+ "grad_norm": 0.06016629561781883,
+ "learning_rate": 0.00039223899519496723,
+ "loss": 2.092460870742798,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.0436374768614769,
+ "learning_rate": 0.00039198469116339757,
+ "loss": 1.5640926361083984,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 0.6233252286911011,
+ "eval_runtime": 40.7412,
+ "eval_samples_per_second": 59.939,
+ "eval_steps_per_second": 0.491,
+ "step": 1512
+ },
+ {
+ "epoch": 21.013937282229964,
+ "grad_norm": 0.05455276742577553,
+ "learning_rate": 0.0003917303154384694,
+ "loss": 2.047119140625,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027874564459932,
+ "grad_norm": 0.07315972447395325,
+ "learning_rate": 0.00039147586822512885,
+ "loss": 2.0495824813842773,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041811846689896,
+ "grad_norm": 0.06641783565282822,
+ "learning_rate": 0.0003912213497283793,
+ "loss": 2.0614805221557617,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05574912891986,
+ "grad_norm": 0.05663100257515907,
+ "learning_rate": 0.0003909667601532819,
+ "loss": 2.042809247970581,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069686411149824,
+ "grad_norm": 0.06034805625677109,
+ "learning_rate": 0.00039071209970495445,
+ "loss": 2.0449604988098145,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083623693379792,
+ "grad_norm": 0.0593884214758873,
+ "learning_rate": 0.00039045736858857273,
+ "loss": 2.041276693344116,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097560975609756,
+ "grad_norm": 0.058692675083875656,
+ "learning_rate": 0.0003902025670093687,
+ "loss": 2.0527422428131104,
+ "step": 1519
+ },
+ {
+ "epoch": 21.11149825783972,
+ "grad_norm": 0.06060221418738365,
+ "learning_rate": 0.00038994769517263156,
+ "loss": 2.0596261024475098,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125435540069688,
+ "grad_norm": 0.06678973138332367,
+ "learning_rate": 0.0003896927532837069,
+ "loss": 2.050360918045044,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139372822299652,
+ "grad_norm": 0.07652897387742996,
+ "learning_rate": 0.0003894377415479967,
+ "loss": 2.0567336082458496,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153310104529616,
+ "grad_norm": 0.08250173926353455,
+ "learning_rate": 0.00038918266017095936,
+ "loss": 2.0620968341827393,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16724738675958,
+ "grad_norm": 0.07214570045471191,
+ "learning_rate": 0.0003889275093581096,
+ "loss": 2.0630033016204834,
+ "step": 1524
+ },
+ {
+ "epoch": 21.181184668989548,
+ "grad_norm": 0.06539473682641983,
+ "learning_rate": 0.00038867228931501737,
+ "loss": 2.0676469802856445,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195121951219512,
+ "grad_norm": 0.06375899165868759,
+ "learning_rate": 0.0003884170002473091,
+ "loss": 2.057408332824707,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209059233449477,
+ "grad_norm": 0.06001587584614754,
+ "learning_rate": 0.0003881616423606666,
+ "loss": 2.041994094848633,
+ "step": 1527
+ },
+ {
+ "epoch": 21.222996515679444,
+ "grad_norm": 0.06002265587449074,
+ "learning_rate": 0.00038790621586082697,
+ "loss": 2.050551414489746,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23693379790941,
+ "grad_norm": 0.06519588828086853,
+ "learning_rate": 0.0003876507209535829,
+ "loss": 2.0467751026153564,
+ "step": 1529
+ },
+ {
+ "epoch": 21.250871080139373,
+ "grad_norm": 0.05902981013059616,
+ "learning_rate": 0.00038739515784478186,
+ "loss": 2.044522285461426,
+ "step": 1530
+ },
+ {
+ "epoch": 21.264808362369337,
+ "grad_norm": 0.06421370059251785,
+ "learning_rate": 0.00038713952674032654,
+ "loss": 2.0658581256866455,
+ "step": 1531
+ },
+ {
+ "epoch": 21.278745644599304,
+ "grad_norm": 0.0618787556886673,
+ "learning_rate": 0.0003868838278461743,
+ "loss": 2.0473392009735107,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29268292682927,
+ "grad_norm": 0.06242169067263603,
+ "learning_rate": 0.0003866280613683371,
+ "loss": 2.0579845905303955,
+ "step": 1533
+ },
+ {
+ "epoch": 21.306620209059233,
+ "grad_norm": 0.06324201077222824,
+ "learning_rate": 0.00038637222751288126,
+ "loss": 2.0529446601867676,
+ "step": 1534
+ },
+ {
+ "epoch": 21.320557491289197,
+ "grad_norm": 0.06349490582942963,
+ "learning_rate": 0.0003861163264859277,
+ "loss": 2.0566954612731934,
+ "step": 1535
+ },
+ {
+ "epoch": 21.334494773519165,
+ "grad_norm": 0.06262438744306564,
+ "learning_rate": 0.0003858603584936511,
+ "loss": 2.058950901031494,
+ "step": 1536
+ },
+ {
+ "epoch": 21.34843205574913,
+ "grad_norm": 0.06206792593002319,
+ "learning_rate": 0.0003856043237422803,
+ "loss": 2.0710246562957764,
+ "step": 1537
+ },
+ {
+ "epoch": 21.362369337979093,
+ "grad_norm": 0.06297925114631653,
+ "learning_rate": 0.0003853482224380978,
+ "loss": 2.0621535778045654,
+ "step": 1538
+ },
+ {
+ "epoch": 21.37630662020906,
+ "grad_norm": 0.06135738268494606,
+ "learning_rate": 0.00038509205478744,
+ "loss": 2.0474977493286133,
+ "step": 1539
+ },
+ {
+ "epoch": 21.390243902439025,
+ "grad_norm": 0.056904129683971405,
+ "learning_rate": 0.0003848358209966963,
+ "loss": 2.063415765762329,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40418118466899,
+ "grad_norm": 0.057480588555336,
+ "learning_rate": 0.0003845795212723099,
+ "loss": 2.067509174346924,
+ "step": 1541
+ },
+ {
+ "epoch": 21.418118466898953,
+ "grad_norm": 0.05868278443813324,
+ "learning_rate": 0.00038432315582077664,
+ "loss": 2.0701723098754883,
+ "step": 1542
+ },
+ {
+ "epoch": 21.43205574912892,
+ "grad_norm": 0.05987158045172691,
+ "learning_rate": 0.0003840667248486456,
+ "loss": 2.061171054840088,
+ "step": 1543
+ },
+ {
+ "epoch": 21.445993031358885,
+ "grad_norm": 0.06049453467130661,
+ "learning_rate": 0.0003838102285625186,
+ "loss": 2.0646986961364746,
+ "step": 1544
+ },
+ {
+ "epoch": 21.45993031358885,
+ "grad_norm": 0.05716792121529579,
+ "learning_rate": 0.00038355366716905006,
+ "loss": 2.0667338371276855,
+ "step": 1545
+ },
+ {
+ "epoch": 21.473867595818817,
+ "grad_norm": 0.05772645026445389,
+ "learning_rate": 0.0003832970408749467,
+ "loss": 2.0723180770874023,
+ "step": 1546
+ },
+ {
+ "epoch": 21.48780487804878,
+ "grad_norm": 0.061401285231113434,
+ "learning_rate": 0.00038304034988696816,
+ "loss": 2.0792477130889893,
+ "step": 1547
+ },
+ {
+ "epoch": 21.501742160278745,
+ "grad_norm": 0.06143321841955185,
+ "learning_rate": 0.00038278359441192516,
+ "loss": 2.059917449951172,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51567944250871,
+ "grad_norm": 0.0586276575922966,
+ "learning_rate": 0.00038252677465668136,
+ "loss": 2.064894199371338,
+ "step": 1549
+ },
+ {
+ "epoch": 21.529616724738677,
+ "grad_norm": 0.05764732509851456,
+ "learning_rate": 0.00038226989082815156,
+ "loss": 2.0622854232788086,
+ "step": 1550
+ },
+ {
+ "epoch": 21.54355400696864,
+ "grad_norm": 0.056712232530117035,
+ "learning_rate": 0.00038201294313330245,
+ "loss": 2.06588077545166,
+ "step": 1551
+ },
+ {
+ "epoch": 21.557491289198605,
+ "grad_norm": 0.056804969906806946,
+ "learning_rate": 0.00038175593177915227,
+ "loss": 2.069946765899658,
+ "step": 1552
+ },
+ {
+ "epoch": 21.571428571428573,
+ "grad_norm": 0.05951021611690521,
+ "learning_rate": 0.0003814988569727704,
+ "loss": 2.0718302726745605,
+ "step": 1553
+ },
+ {
+ "epoch": 21.585365853658537,
+ "grad_norm": 0.05590548366308212,
+ "learning_rate": 0.0003812417189212773,
+ "loss": 2.0757155418395996,
+ "step": 1554
+ },
+ {
+ "epoch": 21.5993031358885,
+ "grad_norm": 0.056640397757291794,
+ "learning_rate": 0.00038098451783184447,
+ "loss": 2.0682835578918457,
+ "step": 1555
+ },
+ {
+ "epoch": 21.613240418118465,
+ "grad_norm": 0.05768130347132683,
+ "learning_rate": 0.00038072725391169416,
+ "loss": 2.073762893676758,
+ "step": 1556
+ },
+ {
+ "epoch": 21.627177700348433,
+ "grad_norm": 0.0627785325050354,
+ "learning_rate": 0.0003804699273680994,
+ "loss": 2.071108341217041,
+ "step": 1557
+ },
+ {
+ "epoch": 21.641114982578397,
+ "grad_norm": 0.06194758415222168,
+ "learning_rate": 0.0003802125384083834,
+ "loss": 2.0614805221557617,
+ "step": 1558
+ },
+ {
+ "epoch": 21.65505226480836,
+ "grad_norm": 0.06017371267080307,
+ "learning_rate": 0.0003799550872399197,
+ "loss": 2.087857246398926,
+ "step": 1559
+ },
+ {
+ "epoch": 21.66898954703833,
+ "grad_norm": 0.057792097330093384,
+ "learning_rate": 0.0003796975740701323,
+ "loss": 2.0456385612487793,
+ "step": 1560
+ },
+ {
+ "epoch": 21.682926829268293,
+ "grad_norm": 0.056328751146793365,
+ "learning_rate": 0.0003794399991064948,
+ "loss": 2.0565314292907715,
+ "step": 1561
+ },
+ {
+ "epoch": 21.696864111498257,
+ "grad_norm": 0.057881973683834076,
+ "learning_rate": 0.00037918236255653074,
+ "loss": 2.073568344116211,
+ "step": 1562
+ },
+ {
+ "epoch": 21.71080139372822,
+ "grad_norm": 0.057316653430461884,
+ "learning_rate": 0.00037892466462781306,
+ "loss": 2.069281578063965,
+ "step": 1563
+ },
+ {
+ "epoch": 21.72473867595819,
+ "grad_norm": 0.059721436351537704,
+ "learning_rate": 0.00037866690552796446,
+ "loss": 2.071974992752075,
+ "step": 1564
+ },
+ {
+ "epoch": 21.738675958188153,
+ "grad_norm": 0.056571077555418015,
+ "learning_rate": 0.00037840908546465696,
+ "loss": 2.073861837387085,
+ "step": 1565
+ },
+ {
+ "epoch": 21.752613240418118,
+ "grad_norm": 0.05739707872271538,
+ "learning_rate": 0.0003781512046456111,
+ "loss": 2.066429376602173,
+ "step": 1566
+ },
+ {
+ "epoch": 21.766550522648085,
+ "grad_norm": 0.06225696951150894,
+ "learning_rate": 0.0003778932632785972,
+ "loss": 2.067384719848633,
+ "step": 1567
+ },
+ {
+ "epoch": 21.78048780487805,
+ "grad_norm": 0.06119474023580551,
+ "learning_rate": 0.00037763526157143376,
+ "loss": 2.066279888153076,
+ "step": 1568
+ },
+ {
+ "epoch": 21.794425087108014,
+ "grad_norm": 0.05468257516622543,
+ "learning_rate": 0.00037737719973198825,
+ "loss": 2.0616049766540527,
+ "step": 1569
+ },
+ {
+ "epoch": 21.808362369337978,
+ "grad_norm": 0.05737150460481644,
+ "learning_rate": 0.0003771190779681762,
+ "loss": 2.084932804107666,
+ "step": 1570
+ },
+ {
+ "epoch": 21.822299651567945,
+ "grad_norm": 0.06001964956521988,
+ "learning_rate": 0.0003768608964879619,
+ "loss": 2.076648712158203,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83623693379791,
+ "grad_norm": 0.05931064859032631,
+ "learning_rate": 0.00037660265549935724,
+ "loss": 2.0765013694763184,
+ "step": 1572
+ },
+ {
+ "epoch": 21.850174216027874,
+ "grad_norm": 0.05849326401948929,
+ "learning_rate": 0.00037634435521042257,
+ "loss": 2.093994617462158,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86411149825784,
+ "grad_norm": 0.05817103758454323,
+ "learning_rate": 0.0003760859958292656,
+ "loss": 2.0881946086883545,
+ "step": 1574
+ },
+ {
+ "epoch": 21.878048780487806,
+ "grad_norm": 0.059196434915065765,
+ "learning_rate": 0.000375827577564042,
+ "loss": 2.0638458728790283,
+ "step": 1575
+ },
+ {
+ "epoch": 21.89198606271777,
+ "grad_norm": 0.059970878064632416,
+ "learning_rate": 0.0003755691006229545,
+ "loss": 2.0775275230407715,
+ "step": 1576
+ },
+ {
+ "epoch": 21.905923344947734,
+ "grad_norm": 0.059128474444150925,
+ "learning_rate": 0.0003753105652142534,
+ "loss": 2.090787172317505,
+ "step": 1577
+ },
+ {
+ "epoch": 21.9198606271777,
+ "grad_norm": 0.060141123831272125,
+ "learning_rate": 0.000375051971546236,
+ "loss": 2.096325159072876,
+ "step": 1578
+ },
+ {
+ "epoch": 21.933797909407666,
+ "grad_norm": 0.05767657607793808,
+ "learning_rate": 0.0003747933198272465,
+ "loss": 2.0847182273864746,
+ "step": 1579
+ },
+ {
+ "epoch": 21.94773519163763,
+ "grad_norm": 0.05725790187716484,
+ "learning_rate": 0.00037453461026567604,
+ "loss": 2.0854201316833496,
+ "step": 1580
+ },
+ {
+ "epoch": 21.961672473867594,
+ "grad_norm": 0.05683537945151329,
+ "learning_rate": 0.00037427584306996196,
+ "loss": 2.0740489959716797,
+ "step": 1581
+ },
+ {
+ "epoch": 21.975609756097562,
+ "grad_norm": 0.0595269575715065,
+ "learning_rate": 0.0003740170184485888,
+ "loss": 2.074122905731201,
+ "step": 1582
+ },
+ {
+ "epoch": 21.989547038327526,
+ "grad_norm": 0.05857275053858757,
+ "learning_rate": 0.0003737581366100864,
+ "loss": 2.078317642211914,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.044675085693597794,
+ "learning_rate": 0.0003734991977630315,
+ "loss": 1.5551567077636719,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 0.6246243715286255,
+ "eval_runtime": 40.365,
+ "eval_samples_per_second": 60.498,
+ "eval_steps_per_second": 0.495,
+ "step": 1584
+ },
+ {
+ "epoch": 22.013937282229964,
+ "grad_norm": 0.05527259409427643,
+ "learning_rate": 0.0003732402021160463,
+ "loss": 2.0354390144348145,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027874564459932,
+ "grad_norm": 0.05814537778496742,
+ "learning_rate": 0.00037298114987779885,
+ "loss": 2.0332770347595215,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041811846689896,
+ "grad_norm": 0.057016633450984955,
+ "learning_rate": 0.000372722041257003,
+ "loss": 2.0279831886291504,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05574912891986,
+ "grad_norm": 0.05586880445480347,
+ "learning_rate": 0.00037246287646241783,
+ "loss": 2.0343117713928223,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069686411149824,
+ "grad_norm": 0.05358196422457695,
+ "learning_rate": 0.0003722036557028478,
+ "loss": 2.0269603729248047,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083623693379792,
+ "grad_norm": 0.05718434974551201,
+ "learning_rate": 0.0003719443791871422,
+ "loss": 2.0179548263549805,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097560975609756,
+ "grad_norm": 0.053914040327072144,
+ "learning_rate": 0.0003716850471241958,
+ "loss": 2.0334277153015137,
+ "step": 1591
+ },
+ {
+ "epoch": 22.11149825783972,
+ "grad_norm": 0.05748381093144417,
+ "learning_rate": 0.0003714256597229474,
+ "loss": 2.0396666526794434,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125435540069688,
+ "grad_norm": 0.056134182959795,
+ "learning_rate": 0.0003711662171923809,
+ "loss": 2.0390894412994385,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139372822299652,
+ "grad_norm": 0.0571996234357357,
+ "learning_rate": 0.0003709067197415244,
+ "loss": 2.0275979042053223,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153310104529616,
+ "grad_norm": 0.061493515968322754,
+ "learning_rate": 0.00037064716757945036,
+ "loss": 2.0350265502929688,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16724738675958,
+ "grad_norm": 0.06279074400663376,
+ "learning_rate": 0.0003703875609152753,
+ "loss": 2.032461166381836,
+ "step": 1596
+ },
+ {
+ "epoch": 22.181184668989548,
+ "grad_norm": 0.06457336246967316,
+ "learning_rate": 0.0003701278999581595,
+ "loss": 2.0221176147460938,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195121951219512,
+ "grad_norm": 0.06783223152160645,
+ "learning_rate": 0.0003698681849173073,
+ "loss": 2.040794610977173,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209059233449477,
+ "grad_norm": 0.059111304581165314,
+ "learning_rate": 0.0003696084160019662,
+ "loss": 2.027994155883789,
+ "step": 1599
+ },
+ {
+ "epoch": 22.222996515679444,
+ "grad_norm": 0.057050321251153946,
+ "learning_rate": 0.0003693485934214274,
+ "loss": 2.0372071266174316,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23693379790941,
+ "grad_norm": 0.06138737127184868,
+ "learning_rate": 0.0003690887173850253,
+ "loss": 2.024506092071533,
+ "step": 1601
+ },
+ {
+ "epoch": 22.250871080139373,
+ "grad_norm": 0.0670398697257042,
+ "learning_rate": 0.0003688287881021374,
+ "loss": 2.049504280090332,
+ "step": 1602
+ },
+ {
+ "epoch": 22.264808362369337,
+ "grad_norm": 0.0606127567589283,
+ "learning_rate": 0.00036856880578218376,
+ "loss": 2.0285115242004395,
+ "step": 1603
+ },
+ {
+ "epoch": 22.278745644599304,
+ "grad_norm": 0.061918050050735474,
+ "learning_rate": 0.00036830877063462786,
+ "loss": 2.039006233215332,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29268292682927,
+ "grad_norm": 0.05838153138756752,
+ "learning_rate": 0.0003680486828689749,
+ "loss": 2.036097288131714,
+ "step": 1605
+ },
+ {
+ "epoch": 22.306620209059233,
+ "grad_norm": 0.05675302818417549,
+ "learning_rate": 0.00036778854269477315,
+ "loss": 2.0531017780303955,
+ "step": 1606
+ },
+ {
+ "epoch": 22.320557491289197,
+ "grad_norm": 0.059629760682582855,
+ "learning_rate": 0.0003675283503216127,
+ "loss": 2.02315616607666,
+ "step": 1607
+ },
+ {
+ "epoch": 22.334494773519165,
+ "grad_norm": 0.06353911012411118,
+ "learning_rate": 0.000367268105959126,
+ "loss": 2.0464940071105957,
+ "step": 1608
+ },
+ {
+ "epoch": 22.34843205574913,
+ "grad_norm": 0.06483582407236099,
+ "learning_rate": 0.00036700780981698705,
+ "loss": 2.042282819747925,
+ "step": 1609
+ },
+ {
+ "epoch": 22.362369337979093,
+ "grad_norm": 0.05890869349241257,
+ "learning_rate": 0.0003667474621049119,
+ "loss": 2.04038143157959,
+ "step": 1610
+ },
+ {
+ "epoch": 22.37630662020906,
+ "grad_norm": 0.059999171644449234,
+ "learning_rate": 0.00036648706303265795,
+ "loss": 2.048435688018799,
+ "step": 1611
+ },
+ {
+ "epoch": 22.390243902439025,
+ "grad_norm": 0.058220699429512024,
+ "learning_rate": 0.000366226612810024,
+ "loss": 2.029154062271118,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40418118466899,
+ "grad_norm": 0.05856891721487045,
+ "learning_rate": 0.00036596611164685025,
+ "loss": 2.0406973361968994,
+ "step": 1613
+ },
+ {
+ "epoch": 22.418118466898953,
+ "grad_norm": 0.05844642221927643,
+ "learning_rate": 0.0003657055597530175,
+ "loss": 2.036428451538086,
+ "step": 1614
+ },
+ {
+ "epoch": 22.43205574912892,
+ "grad_norm": 0.06196592375636101,
+ "learning_rate": 0.00036544495733844797,
+ "loss": 2.042191743850708,
+ "step": 1615
+ },
+ {
+ "epoch": 22.445993031358885,
+ "grad_norm": 0.062180936336517334,
+ "learning_rate": 0.000365184304613104,
+ "loss": 2.047034502029419,
+ "step": 1616
+ },
+ {
+ "epoch": 22.45993031358885,
+ "grad_norm": 0.06386494636535645,
+ "learning_rate": 0.0003649236017869892,
+ "loss": 2.045753002166748,
+ "step": 1617
+ },
+ {
+ "epoch": 22.473867595818817,
+ "grad_norm": 0.06582126021385193,
+ "learning_rate": 0.00036466284907014685,
+ "loss": 2.0422871112823486,
+ "step": 1618
+ },
+ {
+ "epoch": 22.48780487804878,
+ "grad_norm": 0.06352008134126663,
+ "learning_rate": 0.0003644020466726608,
+ "loss": 2.045677661895752,
+ "step": 1619
+ },
+ {
+ "epoch": 22.501742160278745,
+ "grad_norm": 0.05925827845931053,
+ "learning_rate": 0.00036414119480465473,
+ "loss": 2.0356497764587402,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51567944250871,
+ "grad_norm": 0.06234173849225044,
+ "learning_rate": 0.0003638802936762925,
+ "loss": 2.0444488525390625,
+ "step": 1621
+ },
+ {
+ "epoch": 22.529616724738677,
+ "grad_norm": 0.061812568455934525,
+ "learning_rate": 0.0003636193434977771,
+ "loss": 2.0396502017974854,
+ "step": 1622
+ },
+ {
+ "epoch": 22.54355400696864,
+ "grad_norm": 0.060903795063495636,
+ "learning_rate": 0.0003633583444793516,
+ "loss": 2.0547094345092773,
+ "step": 1623
+ },
+ {
+ "epoch": 22.557491289198605,
+ "grad_norm": 0.06328130513429642,
+ "learning_rate": 0.00036309729683129814,
+ "loss": 2.0523388385772705,
+ "step": 1624
+ },
+ {
+ "epoch": 22.571428571428573,
+ "grad_norm": 0.06148819997906685,
+ "learning_rate": 0.00036283620076393814,
+ "loss": 2.0388646125793457,
+ "step": 1625
+ },
+ {
+ "epoch": 22.585365853658537,
+ "grad_norm": 0.06394615024328232,
+ "learning_rate": 0.000362575056487632,
+ "loss": 2.047064781188965,
+ "step": 1626
+ },
+ {
+ "epoch": 22.5993031358885,
+ "grad_norm": 0.06446268409490585,
+ "learning_rate": 0.00036231386421277877,
+ "loss": 2.05059552192688,
+ "step": 1627
+ },
+ {
+ "epoch": 22.613240418118465,
+ "grad_norm": 0.060211826115846634,
+ "learning_rate": 0.0003620526241498167,
+ "loss": 2.0411198139190674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.627177700348433,
+ "grad_norm": 0.0628863275051117,
+ "learning_rate": 0.0003617913365092218,
+ "loss": 2.0659101009368896,
+ "step": 1629
+ },
+ {
+ "epoch": 22.641114982578397,
+ "grad_norm": 0.06347884237766266,
+ "learning_rate": 0.0003615300015015091,
+ "loss": 2.04814076423645,
+ "step": 1630
+ },
+ {
+ "epoch": 22.65505226480836,
+ "grad_norm": 0.06044246256351471,
+ "learning_rate": 0.0003612686193372312,
+ "loss": 2.0628280639648438,
+ "step": 1631
+ },
+ {
+ "epoch": 22.66898954703833,
+ "grad_norm": 0.05990232527256012,
+ "learning_rate": 0.00036100719022697924,
+ "loss": 2.0414810180664062,
+ "step": 1632
+ },
+ {
+ "epoch": 22.682926829268293,
+ "grad_norm": 0.06174647808074951,
+ "learning_rate": 0.0003607457143813819,
+ "loss": 2.0640687942504883,
+ "step": 1633
+ },
+ {
+ "epoch": 22.696864111498257,
+ "grad_norm": 0.05931185558438301,
+ "learning_rate": 0.00036048419201110553,
+ "loss": 2.0781593322753906,
+ "step": 1634
+ },
+ {
+ "epoch": 22.71080139372822,
+ "grad_norm": 0.05911809206008911,
+ "learning_rate": 0.00036022262332685383,
+ "loss": 2.067699432373047,
+ "step": 1635
+ },
+ {
+ "epoch": 22.72473867595819,
+ "grad_norm": 0.05750560015439987,
+ "learning_rate": 0.0003599610085393681,
+ "loss": 2.046065330505371,
+ "step": 1636
+ },
+ {
+ "epoch": 22.738675958188153,
+ "grad_norm": 0.057942718267440796,
+ "learning_rate": 0.0003596993478594267,
+ "loss": 2.041708469390869,
+ "step": 1637
+ },
+ {
+ "epoch": 22.752613240418118,
+ "grad_norm": 0.06379833817481995,
+ "learning_rate": 0.0003594376414978447,
+ "loss": 2.056877613067627,
+ "step": 1638
+ },
+ {
+ "epoch": 22.766550522648085,
+ "grad_norm": 0.06129376217722893,
+ "learning_rate": 0.0003591758896654745,
+ "loss": 2.051968574523926,
+ "step": 1639
+ },
+ {
+ "epoch": 22.78048780487805,
+ "grad_norm": 0.058218032121658325,
+ "learning_rate": 0.0003589140925732045,
+ "loss": 2.073955535888672,
+ "step": 1640
+ },
+ {
+ "epoch": 22.794425087108014,
+ "grad_norm": 0.05933908373117447,
+ "learning_rate": 0.0003586522504319602,
+ "loss": 2.0543787479400635,
+ "step": 1641
+ },
+ {
+ "epoch": 22.808362369337978,
+ "grad_norm": 0.06085696443915367,
+ "learning_rate": 0.0003583903634527029,
+ "loss": 2.050595760345459,
+ "step": 1642
+ },
+ {
+ "epoch": 22.822299651567945,
+ "grad_norm": 0.05955731123685837,
+ "learning_rate": 0.0003581284318464302,
+ "loss": 2.042296886444092,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83623693379791,
+ "grad_norm": 0.05980071797966957,
+ "learning_rate": 0.00035786645582417564,
+ "loss": 2.0681328773498535,
+ "step": 1644
+ },
+ {
+ "epoch": 22.850174216027874,
+ "grad_norm": 0.06467748433351517,
+ "learning_rate": 0.00035760443559700863,
+ "loss": 2.0636940002441406,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86411149825784,
+ "grad_norm": 0.059384699910879135,
+ "learning_rate": 0.00035734237137603417,
+ "loss": 2.061105251312256,
+ "step": 1646
+ },
+ {
+ "epoch": 22.878048780487806,
+ "grad_norm": 0.0601736456155777,
+ "learning_rate": 0.0003570802633723927,
+ "loss": 2.0704731941223145,
+ "step": 1647
+ },
+ {
+ "epoch": 22.89198606271777,
+ "grad_norm": 0.06421475112438202,
+ "learning_rate": 0.0003568181117972597,
+ "loss": 2.0722408294677734,
+ "step": 1648
+ },
+ {
+ "epoch": 22.905923344947734,
+ "grad_norm": 0.05984245613217354,
+ "learning_rate": 0.00035655591686184626,
+ "loss": 2.055701732635498,
+ "step": 1649
+ },
+ {
+ "epoch": 22.9198606271777,
+ "grad_norm": 0.05827873945236206,
+ "learning_rate": 0.0003562936787773979,
+ "loss": 2.054790735244751,
+ "step": 1650
+ },
+ {
+ "epoch": 22.933797909407666,
+ "grad_norm": 0.06271570920944214,
+ "learning_rate": 0.000356031397755195,
+ "loss": 2.0488545894622803,
+ "step": 1651
+ },
+ {
+ "epoch": 22.94773519163763,
+ "grad_norm": 0.058514442294836044,
+ "learning_rate": 0.00035576907400655303,
+ "loss": 2.0719170570373535,
+ "step": 1652
+ },
+ {
+ "epoch": 22.961672473867594,
+ "grad_norm": 0.06007352098822594,
+ "learning_rate": 0.00035550670774282106,
+ "loss": 2.0626220703125,
+ "step": 1653
+ },
+ {
+ "epoch": 22.975609756097562,
+ "grad_norm": 0.05906634032726288,
+ "learning_rate": 0.0003552442991753831,
+ "loss": 2.056274890899658,
+ "step": 1654
+ },
+ {
+ "epoch": 22.989547038327526,
+ "grad_norm": 0.060033153742551804,
+ "learning_rate": 0.00035498184851565697,
+ "loss": 2.0660805702209473,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.04723877087235451,
+ "learning_rate": 0.0003547193559750944,
+ "loss": 1.53598952293396,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 0.6257458329200745,
+ "eval_runtime": 40.2377,
+ "eval_samples_per_second": 60.689,
+ "eval_steps_per_second": 0.497,
+ "step": 1656
+ },
+ {
+ "epoch": 23.013937282229964,
+ "grad_norm": 0.058115169405937195,
+ "learning_rate": 0.00035445682176518087,
+ "loss": 2.0130319595336914,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027874564459932,
+ "grad_norm": 0.06422477960586548,
+ "learning_rate": 0.00035419424609743546,
+ "loss": 2.01822566986084,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041811846689896,
+ "grad_norm": 0.06179831922054291,
+ "learning_rate": 0.00035393162918341065,
+ "loss": 2.0005578994750977,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05574912891986,
+ "grad_norm": 0.05748692899942398,
+ "learning_rate": 0.00035366897123469213,
+ "loss": 2.0156214237213135,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069686411149824,
+ "grad_norm": 0.060799676924943924,
+ "learning_rate": 0.0003534062724628986,
+ "loss": 2.0284323692321777,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083623693379792,
+ "grad_norm": 0.06141313910484314,
+ "learning_rate": 0.00035314353307968195,
+ "loss": 2.0064034461975098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097560975609756,
+ "grad_norm": 0.05791406333446503,
+ "learning_rate": 0.00035288075329672634,
+ "loss": 2.024519920349121,
+ "step": 1663
+ },
+ {
+ "epoch": 23.11149825783972,
+ "grad_norm": 0.05923449993133545,
+ "learning_rate": 0.0003526179333257487,
+ "loss": 2.01115083694458,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125435540069688,
+ "grad_norm": 0.06012846529483795,
+ "learning_rate": 0.0003523550733784984,
+ "loss": 2.000699758529663,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139372822299652,
+ "grad_norm": 0.06160050258040428,
+ "learning_rate": 0.00035209217366675706,
+ "loss": 2.015692710876465,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153310104529616,
+ "grad_norm": 0.05738912150263786,
+ "learning_rate": 0.000351829234402338,
+ "loss": 2.002467632293701,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16724738675958,
+ "grad_norm": 0.06285665184259415,
+ "learning_rate": 0.0003515662557970867,
+ "loss": 2.018192768096924,
+ "step": 1668
+ },
+ {
+ "epoch": 23.181184668989548,
+ "grad_norm": 0.05908574163913727,
+ "learning_rate": 0.0003513032380628804,
+ "loss": 2.0195794105529785,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195121951219512,
+ "grad_norm": 0.0563330352306366,
+ "learning_rate": 0.0003510401814116277,
+ "loss": 1.9947466850280762,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209059233449477,
+ "grad_norm": 0.0638423040509224,
+ "learning_rate": 0.0003507770860552684,
+ "loss": 2.0242462158203125,
+ "step": 1671
+ },
+ {
+ "epoch": 23.222996515679444,
+ "grad_norm": 0.06179089844226837,
+ "learning_rate": 0.00035051395220577397,
+ "loss": 2.008183479309082,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23693379790941,
+ "grad_norm": 0.06259842962026596,
+ "learning_rate": 0.0003502507800751464,
+ "loss": 2.0423057079315186,
+ "step": 1673
+ },
+ {
+ "epoch": 23.250871080139373,
+ "grad_norm": 0.0649559423327446,
+ "learning_rate": 0.0003499875698754187,
+ "loss": 2.0081734657287598,
+ "step": 1674
+ },
+ {
+ "epoch": 23.264808362369337,
+ "grad_norm": 0.06406015902757645,
+ "learning_rate": 0.00034972432181865467,
+ "loss": 2.0132699012756348,
+ "step": 1675
+ },
+ {
+ "epoch": 23.278745644599304,
+ "grad_norm": 0.062094565480947495,
+ "learning_rate": 0.00034946103611694854,
+ "loss": 2.013676404953003,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29268292682927,
+ "grad_norm": 0.05922264978289604,
+ "learning_rate": 0.00034919771298242477,
+ "loss": 2.0171871185302734,
+ "step": 1677
+ },
+ {
+ "epoch": 23.306620209059233,
+ "grad_norm": 0.05957769602537155,
+ "learning_rate": 0.00034893435262723814,
+ "loss": 2.0160932540893555,
+ "step": 1678
+ },
+ {
+ "epoch": 23.320557491289197,
+ "grad_norm": 0.06196262687444687,
+ "learning_rate": 0.00034867095526357325,
+ "loss": 2.0320985317230225,
+ "step": 1679
+ },
+ {
+ "epoch": 23.334494773519165,
+ "grad_norm": 0.06078353151679039,
+ "learning_rate": 0.0003484075211036446,
+ "loss": 2.019073724746704,
+ "step": 1680
+ },
+ {
+ "epoch": 23.34843205574913,
+ "grad_norm": 0.061823632568120956,
+ "learning_rate": 0.0003481440503596964,
+ "loss": 2.022491931915283,
+ "step": 1681
+ },
+ {
+ "epoch": 23.362369337979093,
+ "grad_norm": 0.05937483161687851,
+ "learning_rate": 0.0003478805432440021,
+ "loss": 2.0177359580993652,
+ "step": 1682
+ },
+ {
+ "epoch": 23.37630662020906,
+ "grad_norm": 0.05713704973459244,
+ "learning_rate": 0.0003476169999688648,
+ "loss": 2.015444278717041,
+ "step": 1683
+ },
+ {
+ "epoch": 23.390243902439025,
+ "grad_norm": 0.05961166322231293,
+ "learning_rate": 0.00034735342074661654,
+ "loss": 2.0229737758636475,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40418118466899,
+ "grad_norm": 0.060035206377506256,
+ "learning_rate": 0.0003470898057896183,
+ "loss": 2.014166831970215,
+ "step": 1685
+ },
+ {
+ "epoch": 23.418118466898953,
+ "grad_norm": 0.057613737881183624,
+ "learning_rate": 0.0003468261553102599,
+ "loss": 2.0328001976013184,
+ "step": 1686
+ },
+ {
+ "epoch": 23.43205574912892,
+ "grad_norm": 0.05975262075662613,
+ "learning_rate": 0.00034656246952095984,
+ "loss": 2.022041082382202,
+ "step": 1687
+ },
+ {
+ "epoch": 23.445993031358885,
+ "grad_norm": 0.061379995197057724,
+ "learning_rate": 0.0003462987486341648,
+ "loss": 2.028597354888916,
+ "step": 1688
+ },
+ {
+ "epoch": 23.45993031358885,
+ "grad_norm": 0.05924280732870102,
+ "learning_rate": 0.00034603499286235006,
+ "loss": 2.0182688236236572,
+ "step": 1689
+ },
+ {
+ "epoch": 23.473867595818817,
+ "grad_norm": 0.06229586526751518,
+ "learning_rate": 0.0003457712024180188,
+ "loss": 2.0340843200683594,
+ "step": 1690
+ },
+ {
+ "epoch": 23.48780487804878,
+ "grad_norm": 0.06069250404834747,
+ "learning_rate": 0.0003455073775137025,
+ "loss": 2.029205083847046,
+ "step": 1691
+ },
+ {
+ "epoch": 23.501742160278745,
+ "grad_norm": 0.06134101003408432,
+ "learning_rate": 0.0003452435183619598,
+ "loss": 2.025721311569214,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51567944250871,
+ "grad_norm": 0.06118883565068245,
+ "learning_rate": 0.0003449796251753773,
+ "loss": 2.0322911739349365,
+ "step": 1693
+ },
+ {
+ "epoch": 23.529616724738677,
+ "grad_norm": 0.06123068556189537,
+ "learning_rate": 0.00034471569816656915,
+ "loss": 2.025190830230713,
+ "step": 1694
+ },
+ {
+ "epoch": 23.54355400696864,
+ "grad_norm": 0.06471903622150421,
+ "learning_rate": 0.00034445173754817646,
+ "loss": 2.049294948577881,
+ "step": 1695
+ },
+ {
+ "epoch": 23.557491289198605,
+ "grad_norm": 0.06318801641464233,
+ "learning_rate": 0.00034418774353286747,
+ "loss": 2.0443224906921387,
+ "step": 1696
+ },
+ {
+ "epoch": 23.571428571428573,
+ "grad_norm": 0.060106538236141205,
+ "learning_rate": 0.0003439237163333375,
+ "loss": 2.043574810028076,
+ "step": 1697
+ },
+ {
+ "epoch": 23.585365853658537,
+ "grad_norm": 0.057663049548864365,
+ "learning_rate": 0.0003436596561623084,
+ "loss": 2.018641710281372,
+ "step": 1698
+ },
+ {
+ "epoch": 23.5993031358885,
+ "grad_norm": 0.06039858236908913,
+ "learning_rate": 0.0003433955632325288,
+ "loss": 2.0226235389709473,
+ "step": 1699
+ },
+ {
+ "epoch": 23.613240418118465,
+ "grad_norm": 0.0627814382314682,
+ "learning_rate": 0.00034313143775677353,
+ "loss": 2.0472609996795654,
+ "step": 1700
+ },
+ {
+ "epoch": 23.627177700348433,
+ "grad_norm": 0.06336840242147446,
+ "learning_rate": 0.00034286727994784367,
+ "loss": 2.0199856758117676,
+ "step": 1701
+ },
+ {
+ "epoch": 23.641114982578397,
+ "grad_norm": 0.06368310749530792,
+ "learning_rate": 0.0003426030900185665,
+ "loss": 2.047079086303711,
+ "step": 1702
+ },
+ {
+ "epoch": 23.65505226480836,
+ "grad_norm": 0.0638064444065094,
+ "learning_rate": 0.0003423388681817949,
+ "loss": 2.030421257019043,
+ "step": 1703
+ },
+ {
+ "epoch": 23.66898954703833,
+ "grad_norm": 0.06360456347465515,
+ "learning_rate": 0.00034207461465040793,
+ "loss": 2.042492389678955,
+ "step": 1704
+ },
+ {
+ "epoch": 23.682926829268293,
+ "grad_norm": 0.06266028434038162,
+ "learning_rate": 0.0003418103296373096,
+ "loss": 2.036344528198242,
+ "step": 1705
+ },
+ {
+ "epoch": 23.696864111498257,
+ "grad_norm": 0.059724386781454086,
+ "learning_rate": 0.00034154601335542964,
+ "loss": 2.0373427867889404,
+ "step": 1706
+ },
+ {
+ "epoch": 23.71080139372822,
+ "grad_norm": 0.06086008623242378,
+ "learning_rate": 0.00034128166601772304,
+ "loss": 2.036290168762207,
+ "step": 1707
+ },
+ {
+ "epoch": 23.72473867595819,
+ "grad_norm": 0.06396593153476715,
+ "learning_rate": 0.0003410172878371695,
+ "loss": 2.0262579917907715,
+ "step": 1708
+ },
+ {
+ "epoch": 23.738675958188153,
+ "grad_norm": 0.06352575868368149,
+ "learning_rate": 0.00034075287902677394,
+ "loss": 2.027799367904663,
+ "step": 1709
+ },
+ {
+ "epoch": 23.752613240418118,
+ "grad_norm": 0.058496296405792236,
+ "learning_rate": 0.0003404884397995655,
+ "loss": 2.0130109786987305,
+ "step": 1710
+ },
+ {
+ "epoch": 23.766550522648085,
+ "grad_norm": 0.06340806186199188,
+ "learning_rate": 0.00034022397036859837,
+ "loss": 2.0470035076141357,
+ "step": 1711
+ },
+ {
+ "epoch": 23.78048780487805,
+ "grad_norm": 0.06579892337322235,
+ "learning_rate": 0.0003399594709469506,
+ "loss": 2.0488643646240234,
+ "step": 1712
+ },
+ {
+ "epoch": 23.794425087108014,
+ "grad_norm": 0.06289743632078171,
+ "learning_rate": 0.00033969494174772465,
+ "loss": 2.0357401371002197,
+ "step": 1713
+ },
+ {
+ "epoch": 23.808362369337978,
+ "grad_norm": 0.06139713153243065,
+ "learning_rate": 0.0003394303829840469,
+ "loss": 2.049866199493408,
+ "step": 1714
+ },
+ {
+ "epoch": 23.822299651567945,
+ "grad_norm": 0.060502659529447556,
+ "learning_rate": 0.00033916579486906764,
+ "loss": 2.052330732345581,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83623693379791,
+ "grad_norm": 0.06220673397183418,
+ "learning_rate": 0.0003389011776159607,
+ "loss": 2.033612012863159,
+ "step": 1716
+ },
+ {
+ "epoch": 23.850174216027874,
+ "grad_norm": 0.06053227186203003,
+ "learning_rate": 0.0003386365314379233,
+ "loss": 2.045104503631592,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86411149825784,
+ "grad_norm": 0.05762678012251854,
+ "learning_rate": 0.0003383718565481763,
+ "loss": 2.043063163757324,
+ "step": 1718
+ },
+ {
+ "epoch": 23.878048780487806,
+ "grad_norm": 0.06120038032531738,
+ "learning_rate": 0.0003381071531599633,
+ "loss": 2.0259132385253906,
+ "step": 1719
+ },
+ {
+ "epoch": 23.89198606271777,
+ "grad_norm": 0.05868235602974892,
+ "learning_rate": 0.00033784242148655115,
+ "loss": 2.0501866340637207,
+ "step": 1720
+ },
+ {
+ "epoch": 23.905923344947734,
+ "grad_norm": 0.06055598333477974,
+ "learning_rate": 0.00033757766174122934,
+ "loss": 2.0424137115478516,
+ "step": 1721
+ },
+ {
+ "epoch": 23.9198606271777,
+ "grad_norm": 0.060600437223911285,
+ "learning_rate": 0.00033731287413731005,
+ "loss": 2.0340218544006348,
+ "step": 1722
+ },
+ {
+ "epoch": 23.933797909407666,
+ "grad_norm": 0.05815065652132034,
+ "learning_rate": 0.0003370480588881278,
+ "loss": 2.0448193550109863,
+ "step": 1723
+ },
+ {
+ "epoch": 23.94773519163763,
+ "grad_norm": 0.06015155836939812,
+ "learning_rate": 0.0003367832162070395,
+ "loss": 2.06162428855896,
+ "step": 1724
+ },
+ {
+ "epoch": 23.961672473867594,
+ "grad_norm": 0.0582960844039917,
+ "learning_rate": 0.000336518346307424,
+ "loss": 2.047044038772583,
+ "step": 1725
+ },
+ {
+ "epoch": 23.975609756097562,
+ "grad_norm": 0.05905769765377045,
+ "learning_rate": 0.0003362534494026824,
+ "loss": 2.0379064083099365,
+ "step": 1726
+ },
+ {
+ "epoch": 23.989547038327526,
+ "grad_norm": 0.05936498939990997,
+ "learning_rate": 0.0003359885257062372,
+ "loss": 2.0339996814727783,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.04747362807393074,
+ "learning_rate": 0.0003357235754315328,
+ "loss": 1.5262415409088135,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 0.6274752020835876,
+ "eval_runtime": 40.9131,
+ "eval_samples_per_second": 59.688,
+ "eval_steps_per_second": 0.489,
+ "step": 1728
+ },
+ {
+ "epoch": 24.013937282229964,
+ "grad_norm": 0.05768436938524246,
+ "learning_rate": 0.0003354585987920345,
+ "loss": 1.9963459968566895,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027874564459932,
+ "grad_norm": 0.06296277791261673,
+ "learning_rate": 0.0003351935960012296,
+ "loss": 2.006284475326538,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041811846689896,
+ "grad_norm": 0.06255233287811279,
+ "learning_rate": 0.0003349285672726259,
+ "loss": 2.0035486221313477,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05574912891986,
+ "grad_norm": 0.0630803257226944,
+ "learning_rate": 0.0003346635128197522,
+ "loss": 1.9809913635253906,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069686411149824,
+ "grad_norm": 0.06342404335737228,
+ "learning_rate": 0.0003343984328561581,
+ "loss": 2.003256320953369,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083623693379792,
+ "grad_norm": 0.06045109033584595,
+ "learning_rate": 0.00033413332759541376,
+ "loss": 2.00066876411438,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097560975609756,
+ "grad_norm": 0.060308802872896194,
+ "learning_rate": 0.0003338681972511098,
+ "loss": 2.0134668350219727,
+ "step": 1735
+ },
+ {
+ "epoch": 24.11149825783972,
+ "grad_norm": 0.05871560424566269,
+ "learning_rate": 0.0003336030420368568,
+ "loss": 1.9902323484420776,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125435540069688,
+ "grad_norm": 0.05714292824268341,
+ "learning_rate": 0.0003333378621662857,
+ "loss": 2.001781702041626,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139372822299652,
+ "grad_norm": 0.05740860477089882,
+ "learning_rate": 0.00033307265785304684,
+ "loss": 1.9846689701080322,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153310104529616,
+ "grad_norm": 0.05828196927905083,
+ "learning_rate": 0.00033280742931081076,
+ "loss": 1.9969019889831543,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16724738675958,
+ "grad_norm": 0.061078913509845734,
+ "learning_rate": 0.00033254217675326737,
+ "loss": 2.0022339820861816,
+ "step": 1740
+ },
+ {
+ "epoch": 24.181184668989548,
+ "grad_norm": 0.061026882380247116,
+ "learning_rate": 0.0003322769003941257,
+ "loss": 1.9952095746994019,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195121951219512,
+ "grad_norm": 0.058424439281225204,
+ "learning_rate": 0.00033201160044711423,
+ "loss": 2.016010284423828,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209059233449477,
+ "grad_norm": 0.06072554364800453,
+ "learning_rate": 0.0003317462771259802,
+ "loss": 2.017601728439331,
+ "step": 1743
+ },
+ {
+ "epoch": 24.222996515679444,
+ "grad_norm": 0.05832260847091675,
+ "learning_rate": 0.00033148093064449006,
+ "loss": 1.9968395233154297,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23693379790941,
+ "grad_norm": 0.05976228043437004,
+ "learning_rate": 0.0003312155612164284,
+ "loss": 1.9928964376449585,
+ "step": 1745
+ },
+ {
+ "epoch": 24.250871080139373,
+ "grad_norm": 0.056493740528821945,
+ "learning_rate": 0.00033095016905559883,
+ "loss": 2.0014901161193848,
+ "step": 1746
+ },
+ {
+ "epoch": 24.264808362369337,
+ "grad_norm": 0.05917078256607056,
+ "learning_rate": 0.0003306847543758227,
+ "loss": 1.9978837966918945,
+ "step": 1747
+ },
+ {
+ "epoch": 24.278745644599304,
+ "grad_norm": 0.06398625671863556,
+ "learning_rate": 0.0003304193173909401,
+ "loss": 1.993621587753296,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29268292682927,
+ "grad_norm": 0.0610954612493515,
+ "learning_rate": 0.00033015385831480873,
+ "loss": 1.9844563007354736,
+ "step": 1749
+ },
+ {
+ "epoch": 24.306620209059233,
+ "grad_norm": 0.06027799844741821,
+ "learning_rate": 0.0003298883773613043,
+ "loss": 2.011767864227295,
+ "step": 1750
+ },
+ {
+ "epoch": 24.320557491289197,
+ "grad_norm": 0.059904955327510834,
+ "learning_rate": 0.0003296228747443197,
+ "loss": 1.9990406036376953,
+ "step": 1751
+ },
+ {
+ "epoch": 24.334494773519165,
+ "grad_norm": 0.06107545644044876,
+ "learning_rate": 0.0003293573506777659,
+ "loss": 2.004054307937622,
+ "step": 1752
+ },
+ {
+ "epoch": 24.34843205574913,
+ "grad_norm": 0.0608295314013958,
+ "learning_rate": 0.00032909180537557076,
+ "loss": 1.9942092895507812,
+ "step": 1753
+ },
+ {
+ "epoch": 24.362369337979093,
+ "grad_norm": 0.062092069536447525,
+ "learning_rate": 0.00032882623905167917,
+ "loss": 2.000969886779785,
+ "step": 1754
+ },
+ {
+ "epoch": 24.37630662020906,
+ "grad_norm": 0.06302552670240402,
+ "learning_rate": 0.00032856065192005335,
+ "loss": 2.011990547180176,
+ "step": 1755
+ },
+ {
+ "epoch": 24.390243902439025,
+ "grad_norm": 0.06241324171423912,
+ "learning_rate": 0.00032829504419467194,
+ "loss": 2.0041298866271973,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40418118466899,
+ "grad_norm": 0.06199032813310623,
+ "learning_rate": 0.0003280294160895303,
+ "loss": 1.9974308013916016,
+ "step": 1757
+ },
+ {
+ "epoch": 24.418118466898953,
+ "grad_norm": 0.0632704645395279,
+ "learning_rate": 0.0003277637678186402,
+ "loss": 1.987051010131836,
+ "step": 1758
+ },
+ {
+ "epoch": 24.43205574912892,
+ "grad_norm": 0.0651235282421112,
+ "learning_rate": 0.00032749809959602973,
+ "loss": 2.018052339553833,
+ "step": 1759
+ },
+ {
+ "epoch": 24.445993031358885,
+ "grad_norm": 0.06040659174323082,
+ "learning_rate": 0.0003272324116357428,
+ "loss": 1.9972727298736572,
+ "step": 1760
+ },
+ {
+ "epoch": 24.45993031358885,
+ "grad_norm": 0.06218579411506653,
+ "learning_rate": 0.0003269667041518395,
+ "loss": 2.004945755004883,
+ "step": 1761
+ },
+ {
+ "epoch": 24.473867595818817,
+ "grad_norm": 0.0625823512673378,
+ "learning_rate": 0.0003267009773583956,
+ "loss": 2.0095300674438477,
+ "step": 1762
+ },
+ {
+ "epoch": 24.48780487804878,
+ "grad_norm": 0.06576213985681534,
+ "learning_rate": 0.00032643523146950234,
+ "loss": 2.0181446075439453,
+ "step": 1763
+ },
+ {
+ "epoch": 24.501742160278745,
+ "grad_norm": 0.06759225577116013,
+ "learning_rate": 0.0003261694666992664,
+ "loss": 2.0160670280456543,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51567944250871,
+ "grad_norm": 0.06381658464670181,
+ "learning_rate": 0.0003259036832618096,
+ "loss": 2.01186466217041,
+ "step": 1765
+ },
+ {
+ "epoch": 24.529616724738677,
+ "grad_norm": 0.06042137369513512,
+ "learning_rate": 0.0003256378813712688,
+ "loss": 2.008847713470459,
+ "step": 1766
+ },
+ {
+ "epoch": 24.54355400696864,
+ "grad_norm": 0.06264014542102814,
+ "learning_rate": 0.000325372061241796,
+ "loss": 2.0178637504577637,
+ "step": 1767
+ },
+ {
+ "epoch": 24.557491289198605,
+ "grad_norm": 0.06327593326568604,
+ "learning_rate": 0.00032510622308755746,
+ "loss": 2.0170915126800537,
+ "step": 1768
+ },
+ {
+ "epoch": 24.571428571428573,
+ "grad_norm": 0.0598188154399395,
+ "learning_rate": 0.00032484036712273424,
+ "loss": 2.0074727535247803,
+ "step": 1769
+ },
+ {
+ "epoch": 24.585365853658537,
+ "grad_norm": 0.061695221811532974,
+ "learning_rate": 0.0003245744935615219,
+ "loss": 2.0167722702026367,
+ "step": 1770
+ },
+ {
+ "epoch": 24.5993031358885,
+ "grad_norm": 0.06180686876177788,
+ "learning_rate": 0.0003243086026181296,
+ "loss": 2.009035110473633,
+ "step": 1771
+ },
+ {
+ "epoch": 24.613240418118465,
+ "grad_norm": 0.06220526620745659,
+ "learning_rate": 0.00032404269450678116,
+ "loss": 2.006099224090576,
+ "step": 1772
+ },
+ {
+ "epoch": 24.627177700348433,
+ "grad_norm": 0.06316518783569336,
+ "learning_rate": 0.00032377676944171375,
+ "loss": 2.016641616821289,
+ "step": 1773
+ },
+ {
+ "epoch": 24.641114982578397,
+ "grad_norm": 0.057946689426898956,
+ "learning_rate": 0.00032351082763717857,
+ "loss": 2.017998218536377,
+ "step": 1774
+ },
+ {
+ "epoch": 24.65505226480836,
+ "grad_norm": 0.05899166688323021,
+ "learning_rate": 0.0003232448693074399,
+ "loss": 2.0204315185546875,
+ "step": 1775
+ },
+ {
+ "epoch": 24.66898954703833,
+ "grad_norm": 0.057426705956459045,
+ "learning_rate": 0.00032297889466677575,
+ "loss": 2.0104336738586426,
+ "step": 1776
+ },
+ {
+ "epoch": 24.682926829268293,
+ "grad_norm": 0.057524628937244415,
+ "learning_rate": 0.000322712903929477,
+ "loss": 2.025178909301758,
+ "step": 1777
+ },
+ {
+ "epoch": 24.696864111498257,
+ "grad_norm": 0.06144624203443527,
+ "learning_rate": 0.0003224468973098477,
+ "loss": 2.017031192779541,
+ "step": 1778
+ },
+ {
+ "epoch": 24.71080139372822,
+ "grad_norm": 0.05941846966743469,
+ "learning_rate": 0.0003221808750222044,
+ "loss": 2.0069994926452637,
+ "step": 1779
+ },
+ {
+ "epoch": 24.72473867595819,
+ "grad_norm": 0.05983397737145424,
+ "learning_rate": 0.0003219148372808766,
+ "loss": 2.001835584640503,
+ "step": 1780
+ },
+ {
+ "epoch": 24.738675958188153,
+ "grad_norm": 0.06125850975513458,
+ "learning_rate": 0.00032164878430020606,
+ "loss": 2.0281715393066406,
+ "step": 1781
+ },
+ {
+ "epoch": 24.752613240418118,
+ "grad_norm": 0.05926364287734032,
+ "learning_rate": 0.00032138271629454684,
+ "loss": 2.021087884902954,
+ "step": 1782
+ },
+ {
+ "epoch": 24.766550522648085,
+ "grad_norm": 0.06282161921262741,
+ "learning_rate": 0.00032111663347826505,
+ "loss": 2.0212583541870117,
+ "step": 1783
+ },
+ {
+ "epoch": 24.78048780487805,
+ "grad_norm": 0.06804831326007843,
+ "learning_rate": 0.00032085053606573896,
+ "loss": 2.0088284015655518,
+ "step": 1784
+ },
+ {
+ "epoch": 24.794425087108014,
+ "grad_norm": 0.06296320259571075,
+ "learning_rate": 0.0003205844242713584,
+ "loss": 2.017361640930176,
+ "step": 1785
+ },
+ {
+ "epoch": 24.808362369337978,
+ "grad_norm": 0.059934668242931366,
+ "learning_rate": 0.0003203182983095248,
+ "loss": 2.0179076194763184,
+ "step": 1786
+ },
+ {
+ "epoch": 24.822299651567945,
+ "grad_norm": 0.06056538224220276,
+ "learning_rate": 0.0003200521583946511,
+ "loss": 2.009819746017456,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83623693379791,
+ "grad_norm": 0.06241334602236748,
+ "learning_rate": 0.0003197860047411613,
+ "loss": 2.018239736557007,
+ "step": 1788
+ },
+ {
+ "epoch": 24.850174216027874,
+ "grad_norm": 0.0617685467004776,
+ "learning_rate": 0.00031951983756349066,
+ "loss": 2.023508071899414,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86411149825784,
+ "grad_norm": 0.059675298631191254,
+ "learning_rate": 0.00031925365707608536,
+ "loss": 2.0312063694000244,
+ "step": 1790
+ },
+ {
+ "epoch": 24.878048780487806,
+ "grad_norm": 0.059450648725032806,
+ "learning_rate": 0.00031898746349340204,
+ "loss": 2.019585609436035,
+ "step": 1791
+ },
+ {
+ "epoch": 24.89198606271777,
+ "grad_norm": 0.05922064185142517,
+ "learning_rate": 0.0003187212570299082,
+ "loss": 2.0189690589904785,
+ "step": 1792
+ },
+ {
+ "epoch": 24.905923344947734,
+ "grad_norm": 0.05977880582213402,
+ "learning_rate": 0.0003184550379000815,
+ "loss": 2.0235962867736816,
+ "step": 1793
+ },
+ {
+ "epoch": 24.9198606271777,
+ "grad_norm": 0.06021494418382645,
+ "learning_rate": 0.00031818880631841005,
+ "loss": 2.003030300140381,
+ "step": 1794
+ },
+ {
+ "epoch": 24.933797909407666,
+ "grad_norm": 0.06297598779201508,
+ "learning_rate": 0.0003179225624993915,
+ "loss": 2.0156664848327637,
+ "step": 1795
+ },
+ {
+ "epoch": 24.94773519163763,
+ "grad_norm": 0.06162844970822334,
+ "learning_rate": 0.0003176563066575341,
+ "loss": 2.0309252738952637,
+ "step": 1796
+ },
+ {
+ "epoch": 24.961672473867594,
+ "grad_norm": 0.05988132581114769,
+ "learning_rate": 0.0003173900390073549,
+ "loss": 2.022279977798462,
+ "step": 1797
+ },
+ {
+ "epoch": 24.975609756097562,
+ "grad_norm": 0.06265146285295486,
+ "learning_rate": 0.0003171237597633813,
+ "loss": 2.026237964630127,
+ "step": 1798
+ },
+ {
+ "epoch": 24.989547038327526,
+ "grad_norm": 0.06297934055328369,
+ "learning_rate": 0.0003168574691401496,
+ "loss": 2.0358262062072754,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.04890676215291023,
+ "learning_rate": 0.0003165911673522053,
+ "loss": 1.5189425945281982,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 0.6287577152252197,
+ "eval_runtime": 39.8508,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 1800
+ },
+ {
+ "epoch": 25.013937282229964,
+ "grad_norm": 0.056210409849882126,
+ "learning_rate": 0.00031632485461410294,
+ "loss": 1.9758458137512207,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027874564459932,
+ "grad_norm": 0.06387963145971298,
+ "learning_rate": 0.00031605853114040583,
+ "loss": 1.9743283987045288,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041811846689896,
+ "grad_norm": 0.059931688010692596,
+ "learning_rate": 0.00031579219714568613,
+ "loss": 1.983224630355835,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05574912891986,
+ "grad_norm": 0.059698957949876785,
+ "learning_rate": 0.0003155258528445242,
+ "loss": 1.9638850688934326,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069686411149824,
+ "grad_norm": 0.057556964457035065,
+ "learning_rate": 0.0003152594984515089,
+ "loss": 1.9888463020324707,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083623693379792,
+ "grad_norm": 0.06250642985105515,
+ "learning_rate": 0.00031499313418123704,
+ "loss": 1.9776880741119385,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097560975609756,
+ "grad_norm": 0.06218443810939789,
+ "learning_rate": 0.0003147267602483138,
+ "loss": 1.9904823303222656,
+ "step": 1807
+ },
+ {
+ "epoch": 25.11149825783972,
+ "grad_norm": 0.06147411838173866,
+ "learning_rate": 0.00031446037686735153,
+ "loss": 1.9881136417388916,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125435540069688,
+ "grad_norm": 0.06080890819430351,
+ "learning_rate": 0.0003141939842529706,
+ "loss": 1.9833264350891113,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139372822299652,
+ "grad_norm": 0.057465583086013794,
+ "learning_rate": 0.0003139275826197987,
+ "loss": 1.9764835834503174,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153310104529616,
+ "grad_norm": 0.061488356441259384,
+ "learning_rate": 0.00031366117218247094,
+ "loss": 1.9807426929473877,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16724738675958,
+ "grad_norm": 0.060737013816833496,
+ "learning_rate": 0.00031339475315562916,
+ "loss": 1.9796113967895508,
+ "step": 1812
+ },
+ {
+ "epoch": 25.181184668989548,
+ "grad_norm": 0.060508858412504196,
+ "learning_rate": 0.00031312832575392267,
+ "loss": 1.9862518310546875,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195121951219512,
+ "grad_norm": 0.06062212213873863,
+ "learning_rate": 0.0003128618901920071,
+ "loss": 1.9855456352233887,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209059233449477,
+ "grad_norm": 0.06143258139491081,
+ "learning_rate": 0.0003125954466845446,
+ "loss": 1.9763805866241455,
+ "step": 1815
+ },
+ {
+ "epoch": 25.222996515679444,
+ "grad_norm": 0.06446489691734314,
+ "learning_rate": 0.00031232899544620404,
+ "loss": 1.9674923419952393,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23693379790941,
+ "grad_norm": 0.06179432198405266,
+ "learning_rate": 0.00031206253669166045,
+ "loss": 1.9792906045913696,
+ "step": 1817
+ },
+ {
+ "epoch": 25.250871080139373,
+ "grad_norm": 0.062219955027103424,
+ "learning_rate": 0.0003117960706355947,
+ "loss": 1.983303189277649,
+ "step": 1818
+ },
+ {
+ "epoch": 25.264808362369337,
+ "grad_norm": 0.060631707310676575,
+ "learning_rate": 0.0003115295974926936,
+ "loss": 1.9859293699264526,
+ "step": 1819
+ },
+ {
+ "epoch": 25.278745644599304,
+ "grad_norm": 0.06329842656850815,
+ "learning_rate": 0.00031126311747765005,
+ "loss": 1.988373041152954,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29268292682927,
+ "grad_norm": 0.06290270388126373,
+ "learning_rate": 0.00031099663080516186,
+ "loss": 1.9817759990692139,
+ "step": 1821
+ },
+ {
+ "epoch": 25.306620209059233,
+ "grad_norm": 0.062234487384557724,
+ "learning_rate": 0.0003107301376899327,
+ "loss": 1.9615445137023926,
+ "step": 1822
+ },
+ {
+ "epoch": 25.320557491289197,
+ "grad_norm": 0.06063779070973396,
+ "learning_rate": 0.00031046363834667123,
+ "loss": 1.9736756086349487,
+ "step": 1823
+ },
+ {
+ "epoch": 25.334494773519165,
+ "grad_norm": 0.06245938315987587,
+ "learning_rate": 0.00031019713299009123,
+ "loss": 1.9932905435562134,
+ "step": 1824
+ },
+ {
+ "epoch": 25.34843205574913,
+ "grad_norm": 0.06380622833967209,
+ "learning_rate": 0.000309930621834911,
+ "loss": 1.9695440530776978,
+ "step": 1825
+ },
+ {
+ "epoch": 25.362369337979093,
+ "grad_norm": 0.0626329705119133,
+ "learning_rate": 0.0003096641050958541,
+ "loss": 1.9820970296859741,
+ "step": 1826
+ },
+ {
+ "epoch": 25.37630662020906,
+ "grad_norm": 0.060844022780656815,
+ "learning_rate": 0.00030939758298764795,
+ "loss": 1.9926050901412964,
+ "step": 1827
+ },
+ {
+ "epoch": 25.390243902439025,
+ "grad_norm": 0.06239933893084526,
+ "learning_rate": 0.00030913105572502483,
+ "loss": 2.001865863800049,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40418118466899,
+ "grad_norm": 0.06491564959287643,
+ "learning_rate": 0.00030886452352272095,
+ "loss": 1.9897377490997314,
+ "step": 1829
+ },
+ {
+ "epoch": 25.418118466898953,
+ "grad_norm": 0.05946580320596695,
+ "learning_rate": 0.00030859798659547624,
+ "loss": 1.973170280456543,
+ "step": 1830
+ },
+ {
+ "epoch": 25.43205574912892,
+ "grad_norm": 0.05930647253990173,
+ "learning_rate": 0.0003083314451580349,
+ "loss": 1.9949123859405518,
+ "step": 1831
+ },
+ {
+ "epoch": 25.445993031358885,
+ "grad_norm": 0.06269685924053192,
+ "learning_rate": 0.0003080648994251445,
+ "loss": 1.9912991523742676,
+ "step": 1832
+ },
+ {
+ "epoch": 25.45993031358885,
+ "grad_norm": 0.0637395903468132,
+ "learning_rate": 0.0003077983496115561,
+ "loss": 1.9869414567947388,
+ "step": 1833
+ },
+ {
+ "epoch": 25.473867595818817,
+ "grad_norm": 0.0628778263926506,
+ "learning_rate": 0.00030753179593202406,
+ "loss": 1.9748289585113525,
+ "step": 1834
+ },
+ {
+ "epoch": 25.48780487804878,
+ "grad_norm": 0.06064455583691597,
+ "learning_rate": 0.0003072652386013059,
+ "loss": 2.008216381072998,
+ "step": 1835
+ },
+ {
+ "epoch": 25.501742160278745,
+ "grad_norm": 0.0655415952205658,
+ "learning_rate": 0.00030699867783416207,
+ "loss": 1.9852811098098755,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51567944250871,
+ "grad_norm": 0.06189798563718796,
+ "learning_rate": 0.0003067321138453557,
+ "loss": 1.983716368675232,
+ "step": 1837
+ },
+ {
+ "epoch": 25.529616724738677,
+ "grad_norm": 0.061454080045223236,
+ "learning_rate": 0.0003064655468496527,
+ "loss": 2.013745069503784,
+ "step": 1838
+ },
+ {
+ "epoch": 25.54355400696864,
+ "grad_norm": 0.06712181121110916,
+ "learning_rate": 0.0003061989770618211,
+ "loss": 1.9856178760528564,
+ "step": 1839
+ },
+ {
+ "epoch": 25.557491289198605,
+ "grad_norm": 0.06364461034536362,
+ "learning_rate": 0.00030593240469663154,
+ "loss": 2.0034379959106445,
+ "step": 1840
+ },
+ {
+ "epoch": 25.571428571428573,
+ "grad_norm": 0.06669855862855911,
+ "learning_rate": 0.0003056658299688563,
+ "loss": 1.9868637323379517,
+ "step": 1841
+ },
+ {
+ "epoch": 25.585365853658537,
+ "grad_norm": 0.06578591465950012,
+ "learning_rate": 0.00030539925309327017,
+ "loss": 1.9998185634613037,
+ "step": 1842
+ },
+ {
+ "epoch": 25.5993031358885,
+ "grad_norm": 0.06305381655693054,
+ "learning_rate": 0.0003051326742846491,
+ "loss": 2.000594139099121,
+ "step": 1843
+ },
+ {
+ "epoch": 25.613240418118465,
+ "grad_norm": 0.061369556933641434,
+ "learning_rate": 0.000304866093757771,
+ "loss": 1.9811584949493408,
+ "step": 1844
+ },
+ {
+ "epoch": 25.627177700348433,
+ "grad_norm": 0.062390975654125214,
+ "learning_rate": 0.00030459951172741474,
+ "loss": 2.0020859241485596,
+ "step": 1845
+ },
+ {
+ "epoch": 25.641114982578397,
+ "grad_norm": 0.06477145850658417,
+ "learning_rate": 0.0003043329284083609,
+ "loss": 2.004000186920166,
+ "step": 1846
+ },
+ {
+ "epoch": 25.65505226480836,
+ "grad_norm": 0.06288585811853409,
+ "learning_rate": 0.00030406634401539053,
+ "loss": 1.9919252395629883,
+ "step": 1847
+ },
+ {
+ "epoch": 25.66898954703833,
+ "grad_norm": 0.05959830805659294,
+ "learning_rate": 0.00030379975876328614,
+ "loss": 1.9860963821411133,
+ "step": 1848
+ },
+ {
+ "epoch": 25.682926829268293,
+ "grad_norm": 0.0596182718873024,
+ "learning_rate": 0.00030353317286683053,
+ "loss": 2.0014610290527344,
+ "step": 1849
+ },
+ {
+ "epoch": 25.696864111498257,
+ "grad_norm": 0.06665527820587158,
+ "learning_rate": 0.0003032665865408073,
+ "loss": 2.0072989463806152,
+ "step": 1850
+ },
+ {
+ "epoch": 25.71080139372822,
+ "grad_norm": 0.0631684735417366,
+ "learning_rate": 0.000303,
+ "loss": 2.0076632499694824,
+ "step": 1851
+ },
+ {
+ "epoch": 25.72473867595819,
+ "grad_norm": 0.06214272975921631,
+ "learning_rate": 0.0003027334134591926,
+ "loss": 2.0143537521362305,
+ "step": 1852
+ },
+ {
+ "epoch": 25.738675958188153,
+ "grad_norm": 0.06422685086727142,
+ "learning_rate": 0.00030246682713316945,
+ "loss": 1.9885808229446411,
+ "step": 1853
+ },
+ {
+ "epoch": 25.752613240418118,
+ "grad_norm": 0.06583613157272339,
+ "learning_rate": 0.0003022002412367138,
+ "loss": 2.0042953491210938,
+ "step": 1854
+ },
+ {
+ "epoch": 25.766550522648085,
+ "grad_norm": 0.06265029311180115,
+ "learning_rate": 0.00030193365598460946,
+ "loss": 2.003157138824463,
+ "step": 1855
+ },
+ {
+ "epoch": 25.78048780487805,
+ "grad_norm": 0.0612902007997036,
+ "learning_rate": 0.0003016670715916391,
+ "loss": 1.9832231998443604,
+ "step": 1856
+ },
+ {
+ "epoch": 25.794425087108014,
+ "grad_norm": 0.0625016912817955,
+ "learning_rate": 0.00030140048827258524,
+ "loss": 2.0012407302856445,
+ "step": 1857
+ },
+ {
+ "epoch": 25.808362369337978,
+ "grad_norm": 0.062357913702726364,
+ "learning_rate": 0.0003011339062422289,
+ "loss": 2.010338306427002,
+ "step": 1858
+ },
+ {
+ "epoch": 25.822299651567945,
+ "grad_norm": 0.06349784880876541,
+ "learning_rate": 0.00030086732571535076,
+ "loss": 2.0105648040771484,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83623693379791,
+ "grad_norm": 0.061822179704904556,
+ "learning_rate": 0.0003006007469067297,
+ "loss": 1.9987019300460815,
+ "step": 1860
+ },
+ {
+ "epoch": 25.850174216027874,
+ "grad_norm": 0.06443580985069275,
+ "learning_rate": 0.0003003341700311436,
+ "loss": 1.994943618774414,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86411149825784,
+ "grad_norm": 0.061674103140830994,
+ "learning_rate": 0.0003000675953033685,
+ "loss": 2.0116522312164307,
+ "step": 1862
+ },
+ {
+ "epoch": 25.878048780487806,
+ "grad_norm": 0.061944279819726944,
+ "learning_rate": 0.0002998010229381789,
+ "loss": 2.012296438217163,
+ "step": 1863
+ },
+ {
+ "epoch": 25.89198606271777,
+ "grad_norm": 0.06270956248044968,
+ "learning_rate": 0.0002995344531503474,
+ "loss": 2.0057339668273926,
+ "step": 1864
+ },
+ {
+ "epoch": 25.905923344947734,
+ "grad_norm": 0.06427973508834839,
+ "learning_rate": 0.0002992678861546442,
+ "loss": 2.0053863525390625,
+ "step": 1865
+ },
+ {
+ "epoch": 25.9198606271777,
+ "grad_norm": 0.0637369304895401,
+ "learning_rate": 0.00029900132216583786,
+ "loss": 2.0130667686462402,
+ "step": 1866
+ },
+ {
+ "epoch": 25.933797909407666,
+ "grad_norm": 0.06291835755109787,
+ "learning_rate": 0.0002987347613986941,
+ "loss": 2.0094404220581055,
+ "step": 1867
+ },
+ {
+ "epoch": 25.94773519163763,
+ "grad_norm": 0.06374962627887726,
+ "learning_rate": 0.0002984682040679759,
+ "loss": 2.0074503421783447,
+ "step": 1868
+ },
+ {
+ "epoch": 25.961672473867594,
+ "grad_norm": 0.06330475956201553,
+ "learning_rate": 0.00029820165038844395,
+ "loss": 1.9999325275421143,
+ "step": 1869
+ },
+ {
+ "epoch": 25.975609756097562,
+ "grad_norm": 0.06077918782830238,
+ "learning_rate": 0.00029793510057485556,
+ "loss": 1.994345784187317,
+ "step": 1870
+ },
+ {
+ "epoch": 25.989547038327526,
+ "grad_norm": 0.06524437665939331,
+ "learning_rate": 0.00029766855484196513,
+ "loss": 2.0210394859313965,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.05064176768064499,
+ "learning_rate": 0.0002974020134045238,
+ "loss": 1.4825201034545898,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 0.6298588514328003,
+ "eval_runtime": 43.1768,
+ "eval_samples_per_second": 56.558,
+ "eval_steps_per_second": 0.463,
+ "step": 1872
+ },
+ {
+ "epoch": 26.013937282229964,
+ "grad_norm": 0.05861504375934601,
+ "learning_rate": 0.00029713547647727904,
+ "loss": 1.9728282690048218,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027874564459932,
+ "grad_norm": 0.06309985369443893,
+ "learning_rate": 0.00029686894427497515,
+ "loss": 1.9638245105743408,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041811846689896,
+ "grad_norm": 0.06438111513853073,
+ "learning_rate": 0.0002966024170123519,
+ "loss": 1.9697308540344238,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05574912891986,
+ "grad_norm": 0.060332708060741425,
+ "learning_rate": 0.0002963358949041459,
+ "loss": 1.9574062824249268,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069686411149824,
+ "grad_norm": 0.061463311314582825,
+ "learning_rate": 0.00029606937816508897,
+ "loss": 1.9469952583312988,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083623693379792,
+ "grad_norm": 0.06337480992078781,
+ "learning_rate": 0.00029580286700990887,
+ "loss": 1.9695435762405396,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097560975609756,
+ "grad_norm": 0.06303273886442184,
+ "learning_rate": 0.00029553636165332876,
+ "loss": 1.9764964580535889,
+ "step": 1879
+ },
+ {
+ "epoch": 26.11149825783972,
+ "grad_norm": 0.06061594933271408,
+ "learning_rate": 0.0002952698623100673,
+ "loss": 1.9758808612823486,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125435540069688,
+ "grad_norm": 0.061181873083114624,
+ "learning_rate": 0.0002950033691948382,
+ "loss": 1.9607892036437988,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139372822299652,
+ "grad_norm": 0.06557024270296097,
+ "learning_rate": 0.00029473688252234994,
+ "loss": 1.9752333164215088,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153310104529616,
+ "grad_norm": 0.06261707097291946,
+ "learning_rate": 0.0002944704025073064,
+ "loss": 1.9634147882461548,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16724738675958,
+ "grad_norm": 0.05993996560573578,
+ "learning_rate": 0.00029420392936440525,
+ "loss": 1.957287311553955,
+ "step": 1884
+ },
+ {
+ "epoch": 26.181184668989548,
+ "grad_norm": 0.0612766332924366,
+ "learning_rate": 0.00029393746330833953,
+ "loss": 1.9562265872955322,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195121951219512,
+ "grad_norm": 0.05735815688967705,
+ "learning_rate": 0.0002936710045537959,
+ "loss": 1.9577300548553467,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209059233449477,
+ "grad_norm": 0.06258611381053925,
+ "learning_rate": 0.0002934045533154554,
+ "loss": 1.9661660194396973,
+ "step": 1887
+ },
+ {
+ "epoch": 26.222996515679444,
+ "grad_norm": 0.06298062205314636,
+ "learning_rate": 0.00029313810980799296,
+ "loss": 1.9508070945739746,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23693379790941,
+ "grad_norm": 0.06346822530031204,
+ "learning_rate": 0.00029287167424607726,
+ "loss": 1.9677343368530273,
+ "step": 1889
+ },
+ {
+ "epoch": 26.250871080139373,
+ "grad_norm": 0.05959425866603851,
+ "learning_rate": 0.00029260524684437077,
+ "loss": 1.9574894905090332,
+ "step": 1890
+ },
+ {
+ "epoch": 26.264808362369337,
+ "grad_norm": 0.0610729418694973,
+ "learning_rate": 0.00029233882781752905,
+ "loss": 1.9515372514724731,
+ "step": 1891
+ },
+ {
+ "epoch": 26.278745644599304,
+ "grad_norm": 0.06362602859735489,
+ "learning_rate": 0.00029207241738020136,
+ "loss": 1.966064453125,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29268292682927,
+ "grad_norm": 0.06124011054635048,
+ "learning_rate": 0.0002918060157470294,
+ "loss": 1.9654541015625,
+ "step": 1893
+ },
+ {
+ "epoch": 26.306620209059233,
+ "grad_norm": 0.05773317068815231,
+ "learning_rate": 0.00029153962313264856,
+ "loss": 1.9610934257507324,
+ "step": 1894
+ },
+ {
+ "epoch": 26.320557491289197,
+ "grad_norm": 0.05896785855293274,
+ "learning_rate": 0.0002912732397516862,
+ "loss": 1.965133547782898,
+ "step": 1895
+ },
+ {
+ "epoch": 26.334494773519165,
+ "grad_norm": 0.059504859149456024,
+ "learning_rate": 0.00029100686581876283,
+ "loss": 1.9617180824279785,
+ "step": 1896
+ },
+ {
+ "epoch": 26.34843205574913,
+ "grad_norm": 0.05804497376084328,
+ "learning_rate": 0.0002907405015484911,
+ "loss": 1.9761323928833008,
+ "step": 1897
+ },
+ {
+ "epoch": 26.362369337979093,
+ "grad_norm": 0.058211296796798706,
+ "learning_rate": 0.0002904741471554758,
+ "loss": 1.9805936813354492,
+ "step": 1898
+ },
+ {
+ "epoch": 26.37630662020906,
+ "grad_norm": 0.05976752191781998,
+ "learning_rate": 0.0002902078028543139,
+ "loss": 1.9514508247375488,
+ "step": 1899
+ },
+ {
+ "epoch": 26.390243902439025,
+ "grad_norm": 0.05824121832847595,
+ "learning_rate": 0.00028994146885959416,
+ "loss": 1.97285795211792,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40418118466899,
+ "grad_norm": 0.05793534219264984,
+ "learning_rate": 0.00028967514538589715,
+ "loss": 1.9608209133148193,
+ "step": 1901
+ },
+ {
+ "epoch": 26.418118466898953,
+ "grad_norm": 0.05697625130414963,
+ "learning_rate": 0.0002894088326477947,
+ "loss": 1.9787089824676514,
+ "step": 1902
+ },
+ {
+ "epoch": 26.43205574912892,
+ "grad_norm": 0.05703677609562874,
+ "learning_rate": 0.0002891425308598503,
+ "loss": 1.9610021114349365,
+ "step": 1903
+ },
+ {
+ "epoch": 26.445993031358885,
+ "grad_norm": 0.057380311191082,
+ "learning_rate": 0.00028887624023661866,
+ "loss": 1.9764081239700317,
+ "step": 1904
+ },
+ {
+ "epoch": 26.45993031358885,
+ "grad_norm": 0.05846763774752617,
+ "learning_rate": 0.00028860996099264505,
+ "loss": 1.9641292095184326,
+ "step": 1905
+ },
+ {
+ "epoch": 26.473867595818817,
+ "grad_norm": 0.05929148569703102,
+ "learning_rate": 0.000288343693342466,
+ "loss": 1.9803684949874878,
+ "step": 1906
+ },
+ {
+ "epoch": 26.48780487804878,
+ "grad_norm": 0.058236971497535706,
+ "learning_rate": 0.00028807743750060847,
+ "loss": 1.981757402420044,
+ "step": 1907
+ },
+ {
+ "epoch": 26.501742160278745,
+ "grad_norm": 0.05845103785395622,
+ "learning_rate": 0.00028781119368159004,
+ "loss": 1.9736125469207764,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51567944250871,
+ "grad_norm": 0.06104041263461113,
+ "learning_rate": 0.00028754496209991843,
+ "loss": 1.9698188304901123,
+ "step": 1909
+ },
+ {
+ "epoch": 26.529616724738677,
+ "grad_norm": 0.0628688707947731,
+ "learning_rate": 0.0002872787429700917,
+ "loss": 1.9786568880081177,
+ "step": 1910
+ },
+ {
+ "epoch": 26.54355400696864,
+ "grad_norm": 0.06115978956222534,
+ "learning_rate": 0.0002870125365065979,
+ "loss": 1.9744799137115479,
+ "step": 1911
+ },
+ {
+ "epoch": 26.557491289198605,
+ "grad_norm": 0.06233362853527069,
+ "learning_rate": 0.00028674634292391463,
+ "loss": 1.9801995754241943,
+ "step": 1912
+ },
+ {
+ "epoch": 26.571428571428573,
+ "grad_norm": 0.06182500347495079,
+ "learning_rate": 0.0002864801624365093,
+ "loss": 1.9888432025909424,
+ "step": 1913
+ },
+ {
+ "epoch": 26.585365853658537,
+ "grad_norm": 0.05962767079472542,
+ "learning_rate": 0.0002862139952588387,
+ "loss": 1.9648957252502441,
+ "step": 1914
+ },
+ {
+ "epoch": 26.5993031358885,
+ "grad_norm": 0.05994252488017082,
+ "learning_rate": 0.000285947841605349,
+ "loss": 1.9709041118621826,
+ "step": 1915
+ },
+ {
+ "epoch": 26.613240418118465,
+ "grad_norm": 0.06045769527554512,
+ "learning_rate": 0.00028568170169047523,
+ "loss": 1.994588017463684,
+ "step": 1916
+ },
+ {
+ "epoch": 26.627177700348433,
+ "grad_norm": 0.059465229511260986,
+ "learning_rate": 0.00028541557572864154,
+ "loss": 1.9743266105651855,
+ "step": 1917
+ },
+ {
+ "epoch": 26.641114982578397,
+ "grad_norm": 0.06020190939307213,
+ "learning_rate": 0.000285149463934261,
+ "loss": 1.9780995845794678,
+ "step": 1918
+ },
+ {
+ "epoch": 26.65505226480836,
+ "grad_norm": 0.05829813703894615,
+ "learning_rate": 0.0002848833665217349,
+ "loss": 1.9831392765045166,
+ "step": 1919
+ },
+ {
+ "epoch": 26.66898954703833,
+ "grad_norm": 0.05960174649953842,
+ "learning_rate": 0.0002846172837054532,
+ "loss": 1.9799871444702148,
+ "step": 1920
+ },
+ {
+ "epoch": 26.682926829268293,
+ "grad_norm": 0.06133116036653519,
+ "learning_rate": 0.0002843512156997939,
+ "loss": 1.985158920288086,
+ "step": 1921
+ },
+ {
+ "epoch": 26.696864111498257,
+ "grad_norm": 0.05947015434503555,
+ "learning_rate": 0.0002840851627191234,
+ "loss": 1.9783923625946045,
+ "step": 1922
+ },
+ {
+ "epoch": 26.71080139372822,
+ "grad_norm": 0.061687882989645004,
+ "learning_rate": 0.0002838191249777955,
+ "loss": 1.9685261249542236,
+ "step": 1923
+ },
+ {
+ "epoch": 26.72473867595819,
+ "grad_norm": 0.0613567978143692,
+ "learning_rate": 0.00028355310269015225,
+ "loss": 1.9945828914642334,
+ "step": 1924
+ },
+ {
+ "epoch": 26.738675958188153,
+ "grad_norm": 0.05902605876326561,
+ "learning_rate": 0.00028328709607052297,
+ "loss": 1.966120719909668,
+ "step": 1925
+ },
+ {
+ "epoch": 26.752613240418118,
+ "grad_norm": 0.06018664687871933,
+ "learning_rate": 0.0002830211053332242,
+ "loss": 1.991615653038025,
+ "step": 1926
+ },
+ {
+ "epoch": 26.766550522648085,
+ "grad_norm": 0.061171986162662506,
+ "learning_rate": 0.0002827551306925601,
+ "loss": 1.9877536296844482,
+ "step": 1927
+ },
+ {
+ "epoch": 26.78048780487805,
+ "grad_norm": 0.06263386458158493,
+ "learning_rate": 0.00028248917236282147,
+ "loss": 1.9799100160598755,
+ "step": 1928
+ },
+ {
+ "epoch": 26.794425087108014,
+ "grad_norm": 0.06326548755168915,
+ "learning_rate": 0.0002822232305582863,
+ "loss": 1.9805757999420166,
+ "step": 1929
+ },
+ {
+ "epoch": 26.808362369337978,
+ "grad_norm": 0.06226228550076485,
+ "learning_rate": 0.0002819573054932188,
+ "loss": 1.9867990016937256,
+ "step": 1930
+ },
+ {
+ "epoch": 26.822299651567945,
+ "grad_norm": 0.06003205478191376,
+ "learning_rate": 0.0002816913973818703,
+ "loss": 1.9632174968719482,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83623693379791,
+ "grad_norm": 0.06075739488005638,
+ "learning_rate": 0.0002814255064384781,
+ "loss": 1.9638378620147705,
+ "step": 1932
+ },
+ {
+ "epoch": 26.850174216027874,
+ "grad_norm": 0.061726443469524384,
+ "learning_rate": 0.00028115963287726563,
+ "loss": 1.9865589141845703,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86411149825784,
+ "grad_norm": 0.0621049702167511,
+ "learning_rate": 0.0002808937769124425,
+ "loss": 2.007071018218994,
+ "step": 1934
+ },
+ {
+ "epoch": 26.878048780487806,
+ "grad_norm": 0.061793744564056396,
+ "learning_rate": 0.000280627938758204,
+ "loss": 1.988924264907837,
+ "step": 1935
+ },
+ {
+ "epoch": 26.89198606271777,
+ "grad_norm": 0.062355563044548035,
+ "learning_rate": 0.00028036211862873117,
+ "loss": 1.9830653667449951,
+ "step": 1936
+ },
+ {
+ "epoch": 26.905923344947734,
+ "grad_norm": 0.060881514102220535,
+ "learning_rate": 0.0002800963167381904,
+ "loss": 2.003551959991455,
+ "step": 1937
+ },
+ {
+ "epoch": 26.9198606271777,
+ "grad_norm": 0.061703141778707504,
+ "learning_rate": 0.00027983053330073355,
+ "loss": 1.9811534881591797,
+ "step": 1938
+ },
+ {
+ "epoch": 26.933797909407666,
+ "grad_norm": 0.0634910985827446,
+ "learning_rate": 0.00027956476853049765,
+ "loss": 1.9840846061706543,
+ "step": 1939
+ },
+ {
+ "epoch": 26.94773519163763,
+ "grad_norm": 0.060807306319475174,
+ "learning_rate": 0.00027929902264160435,
+ "loss": 1.9879541397094727,
+ "step": 1940
+ },
+ {
+ "epoch": 26.961672473867594,
+ "grad_norm": 0.06126011535525322,
+ "learning_rate": 0.00027903329584816046,
+ "loss": 1.993988037109375,
+ "step": 1941
+ },
+ {
+ "epoch": 26.975609756097562,
+ "grad_norm": 0.06394185870885849,
+ "learning_rate": 0.0002787675883642572,
+ "loss": 1.983365535736084,
+ "step": 1942
+ },
+ {
+ "epoch": 26.989547038327526,
+ "grad_norm": 0.06164516508579254,
+ "learning_rate": 0.00027850190040397036,
+ "loss": 1.989990234375,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.04997039958834648,
+ "learning_rate": 0.00027823623218135976,
+ "loss": 1.4921388626098633,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 0.6310494542121887,
+ "eval_runtime": 40.875,
+ "eval_samples_per_second": 59.743,
+ "eval_steps_per_second": 0.489,
+ "step": 1944
+ },
+ {
+ "epoch": 27.013937282229964,
+ "grad_norm": 0.05894223228096962,
+ "learning_rate": 0.0002779705839104697,
+ "loss": 1.935828447341919,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027874564459932,
+ "grad_norm": 0.06731119751930237,
+ "learning_rate": 0.0002777049558053281,
+ "loss": 1.945769190788269,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041811846689896,
+ "grad_norm": 0.06787799298763275,
+ "learning_rate": 0.0002774393480799466,
+ "loss": 1.944061040878296,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05574912891986,
+ "grad_norm": 0.05840576812624931,
+ "learning_rate": 0.00027717376094832076,
+ "loss": 1.94246506690979,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069686411149824,
+ "grad_norm": 0.058151841163635254,
+ "learning_rate": 0.0002769081946244293,
+ "loss": 1.940029263496399,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083623693379792,
+ "grad_norm": 0.06329379975795746,
+ "learning_rate": 0.00027664264932223407,
+ "loss": 1.9523427486419678,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097560975609756,
+ "grad_norm": 0.06530784070491791,
+ "learning_rate": 0.00027637712525568024,
+ "loss": 1.9352362155914307,
+ "step": 1951
+ },
+ {
+ "epoch": 27.11149825783972,
+ "grad_norm": 0.06273649632930756,
+ "learning_rate": 0.0002761116226386958,
+ "loss": 1.946252703666687,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125435540069688,
+ "grad_norm": 0.05996445193886757,
+ "learning_rate": 0.00027584614168519125,
+ "loss": 1.9437150955200195,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139372822299652,
+ "grad_norm": 0.059087123721838,
+ "learning_rate": 0.00027558068260905976,
+ "loss": 1.936135172843933,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153310104529616,
+ "grad_norm": 0.06257616728544235,
+ "learning_rate": 0.00027531524562417734,
+ "loss": 1.9452544450759888,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16724738675958,
+ "grad_norm": 0.06591545045375824,
+ "learning_rate": 0.00027504983094440126,
+ "loss": 1.9334893226623535,
+ "step": 1956
+ },
+ {
+ "epoch": 27.181184668989548,
+ "grad_norm": 0.0595872700214386,
+ "learning_rate": 0.0002747844387835716,
+ "loss": 1.9514997005462646,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195121951219512,
+ "grad_norm": 0.05855429172515869,
+ "learning_rate": 0.0002745190693555099,
+ "loss": 1.955148458480835,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209059233449477,
+ "grad_norm": 0.06383495032787323,
+ "learning_rate": 0.0002742537228740197,
+ "loss": 1.9607688188552856,
+ "step": 1959
+ },
+ {
+ "epoch": 27.222996515679444,
+ "grad_norm": 0.06486682593822479,
+ "learning_rate": 0.00027398839955288575,
+ "loss": 1.941009283065796,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23693379790941,
+ "grad_norm": 0.06293730437755585,
+ "learning_rate": 0.00027372309960587424,
+ "loss": 1.9550879001617432,
+ "step": 1961
+ },
+ {
+ "epoch": 27.250871080139373,
+ "grad_norm": 0.06104227155447006,
+ "learning_rate": 0.0002734578232467327,
+ "loss": 1.9483157396316528,
+ "step": 1962
+ },
+ {
+ "epoch": 27.264808362369337,
+ "grad_norm": 0.06920639425516129,
+ "learning_rate": 0.0002731925706891891,
+ "loss": 1.948688268661499,
+ "step": 1963
+ },
+ {
+ "epoch": 27.278745644599304,
+ "grad_norm": 0.06186635419726372,
+ "learning_rate": 0.0002729273421469532,
+ "loss": 1.9512240886688232,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29268292682927,
+ "grad_norm": 0.06258274614810944,
+ "learning_rate": 0.00027266213783371434,
+ "loss": 1.9538028240203857,
+ "step": 1965
+ },
+ {
+ "epoch": 27.306620209059233,
+ "grad_norm": 0.06546445935964584,
+ "learning_rate": 0.00027239695796314327,
+ "loss": 1.9497663974761963,
+ "step": 1966
+ },
+ {
+ "epoch": 27.320557491289197,
+ "grad_norm": 0.06289096921682358,
+ "learning_rate": 0.0002721318027488902,
+ "loss": 1.9706182479858398,
+ "step": 1967
+ },
+ {
+ "epoch": 27.334494773519165,
+ "grad_norm": 0.06412214040756226,
+ "learning_rate": 0.00027186667240458617,
+ "loss": 1.9423331022262573,
+ "step": 1968
+ },
+ {
+ "epoch": 27.34843205574913,
+ "grad_norm": 0.062773697078228,
+ "learning_rate": 0.0002716015671438419,
+ "loss": 1.9471882581710815,
+ "step": 1969
+ },
+ {
+ "epoch": 27.362369337979093,
+ "grad_norm": 0.060290850698947906,
+ "learning_rate": 0.0002713364871802478,
+ "loss": 1.9732742309570312,
+ "step": 1970
+ },
+ {
+ "epoch": 27.37630662020906,
+ "grad_norm": 0.062357380986213684,
+ "learning_rate": 0.0002710714327273741,
+ "loss": 1.957550048828125,
+ "step": 1971
+ },
+ {
+ "epoch": 27.390243902439025,
+ "grad_norm": 0.06070727854967117,
+ "learning_rate": 0.0002708064039987704,
+ "loss": 1.951438069343567,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40418118466899,
+ "grad_norm": 0.06380726397037506,
+ "learning_rate": 0.00027054140120796546,
+ "loss": 1.9512609243392944,
+ "step": 1973
+ },
+ {
+ "epoch": 27.418118466898953,
+ "grad_norm": 0.06135795637965202,
+ "learning_rate": 0.0002702764245684672,
+ "loss": 1.9663273096084595,
+ "step": 1974
+ },
+ {
+ "epoch": 27.43205574912892,
+ "grad_norm": 0.0627567246556282,
+ "learning_rate": 0.0002700114742937627,
+ "loss": 1.964203119277954,
+ "step": 1975
+ },
+ {
+ "epoch": 27.445993031358885,
+ "grad_norm": 0.06235141307115555,
+ "learning_rate": 0.00026974655059731754,
+ "loss": 1.960247278213501,
+ "step": 1976
+ },
+ {
+ "epoch": 27.45993031358885,
+ "grad_norm": 0.06187676638364792,
+ "learning_rate": 0.0002694816536925759,
+ "loss": 1.9667054414749146,
+ "step": 1977
+ },
+ {
+ "epoch": 27.473867595818817,
+ "grad_norm": 0.06090894341468811,
+ "learning_rate": 0.00026921678379296057,
+ "loss": 1.9695777893066406,
+ "step": 1978
+ },
+ {
+ "epoch": 27.48780487804878,
+ "grad_norm": 0.06363680213689804,
+ "learning_rate": 0.0002689519411118722,
+ "loss": 1.9529361724853516,
+ "step": 1979
+ },
+ {
+ "epoch": 27.501742160278745,
+ "grad_norm": 0.06053898110985756,
+ "learning_rate": 0.00026868712586269,
+ "loss": 1.9707324504852295,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51567944250871,
+ "grad_norm": 0.061363156884908676,
+ "learning_rate": 0.00026842233825877064,
+ "loss": 1.955542802810669,
+ "step": 1981
+ },
+ {
+ "epoch": 27.529616724738677,
+ "grad_norm": 0.06158224865794182,
+ "learning_rate": 0.0002681575785134487,
+ "loss": 1.9489967823028564,
+ "step": 1982
+ },
+ {
+ "epoch": 27.54355400696864,
+ "grad_norm": 0.05886901170015335,
+ "learning_rate": 0.0002678928468400366,
+ "loss": 1.9705257415771484,
+ "step": 1983
+ },
+ {
+ "epoch": 27.557491289198605,
+ "grad_norm": 0.05890485271811485,
+ "learning_rate": 0.00026762814345182367,
+ "loss": 1.9430618286132812,
+ "step": 1984
+ },
+ {
+ "epoch": 27.571428571428573,
+ "grad_norm": 0.060433268547058105,
+ "learning_rate": 0.0002673634685620767,
+ "loss": 1.9634191989898682,
+ "step": 1985
+ },
+ {
+ "epoch": 27.585365853658537,
+ "grad_norm": 0.0626354068517685,
+ "learning_rate": 0.0002670988223840393,
+ "loss": 1.9568289518356323,
+ "step": 1986
+ },
+ {
+ "epoch": 27.5993031358885,
+ "grad_norm": 0.05973317474126816,
+ "learning_rate": 0.0002668342051309324,
+ "loss": 1.9678292274475098,
+ "step": 1987
+ },
+ {
+ "epoch": 27.613240418118465,
+ "grad_norm": 0.062289632856845856,
+ "learning_rate": 0.00026656961701595306,
+ "loss": 1.9413363933563232,
+ "step": 1988
+ },
+ {
+ "epoch": 27.627177700348433,
+ "grad_norm": 0.06162180379033089,
+ "learning_rate": 0.0002663050582522753,
+ "loss": 1.9835751056671143,
+ "step": 1989
+ },
+ {
+ "epoch": 27.641114982578397,
+ "grad_norm": 0.06176188960671425,
+ "learning_rate": 0.00026604052905304946,
+ "loss": 1.9541220664978027,
+ "step": 1990
+ },
+ {
+ "epoch": 27.65505226480836,
+ "grad_norm": 0.06479815393686295,
+ "learning_rate": 0.00026577602963140156,
+ "loss": 1.9596972465515137,
+ "step": 1991
+ },
+ {
+ "epoch": 27.66898954703833,
+ "grad_norm": 0.06571854650974274,
+ "learning_rate": 0.0002655115602004345,
+ "loss": 1.9628980159759521,
+ "step": 1992
+ },
+ {
+ "epoch": 27.682926829268293,
+ "grad_norm": 0.06167568638920784,
+ "learning_rate": 0.00026524712097322604,
+ "loss": 1.9624769687652588,
+ "step": 1993
+ },
+ {
+ "epoch": 27.696864111498257,
+ "grad_norm": 0.06207694113254547,
+ "learning_rate": 0.0002649827121628305,
+ "loss": 1.9487217664718628,
+ "step": 1994
+ },
+ {
+ "epoch": 27.71080139372822,
+ "grad_norm": 0.061430424451828,
+ "learning_rate": 0.00026471833398227694,
+ "loss": 1.9567784070968628,
+ "step": 1995
+ },
+ {
+ "epoch": 27.72473867595819,
+ "grad_norm": 0.06589192897081375,
+ "learning_rate": 0.0002644539866445703,
+ "loss": 1.9594354629516602,
+ "step": 1996
+ },
+ {
+ "epoch": 27.738675958188153,
+ "grad_norm": 0.06640033423900604,
+ "learning_rate": 0.00026418967036269045,
+ "loss": 1.9637835025787354,
+ "step": 1997
+ },
+ {
+ "epoch": 27.752613240418118,
+ "grad_norm": 0.06181980296969414,
+ "learning_rate": 0.0002639253853495921,
+ "loss": 1.9597129821777344,
+ "step": 1998
+ },
+ {
+ "epoch": 27.766550522648085,
+ "grad_norm": 0.06352445483207703,
+ "learning_rate": 0.00026366113181820506,
+ "loss": 1.9612562656402588,
+ "step": 1999
+ },
+ {
+ "epoch": 27.78048780487805,
+ "grad_norm": 0.06398574262857437,
+ "learning_rate": 0.00026339690998143346,
+ "loss": 1.9662930965423584,
+ "step": 2000
+ },
+ {
+ "epoch": 27.794425087108014,
+ "grad_norm": 0.06195084750652313,
+ "learning_rate": 0.00026313272005215637,
+ "loss": 1.956035852432251,
+ "step": 2001
+ },
+ {
+ "epoch": 27.808362369337978,
+ "grad_norm": 0.05951351299881935,
+ "learning_rate": 0.00026286856224322645,
+ "loss": 1.9712438583374023,
+ "step": 2002
+ },
+ {
+ "epoch": 27.822299651567945,
+ "grad_norm": 0.062185633927583694,
+ "learning_rate": 0.00026260443676747114,
+ "loss": 1.9714539051055908,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83623693379791,
+ "grad_norm": 0.061178892850875854,
+ "learning_rate": 0.00026234034383769157,
+ "loss": 1.9691438674926758,
+ "step": 2004
+ },
+ {
+ "epoch": 27.850174216027874,
+ "grad_norm": 0.06262431293725967,
+ "learning_rate": 0.0002620762836666625,
+ "loss": 1.9506231546401978,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86411149825784,
+ "grad_norm": 0.06206081807613373,
+ "learning_rate": 0.00026181225646713257,
+ "loss": 1.9662809371948242,
+ "step": 2006
+ },
+ {
+ "epoch": 27.878048780487806,
+ "grad_norm": 0.0626990869641304,
+ "learning_rate": 0.0002615482624518236,
+ "loss": 1.972696304321289,
+ "step": 2007
+ },
+ {
+ "epoch": 27.89198606271777,
+ "grad_norm": 0.06345485150814056,
+ "learning_rate": 0.0002612843018334309,
+ "loss": 1.9560344219207764,
+ "step": 2008
+ },
+ {
+ "epoch": 27.905923344947734,
+ "grad_norm": 0.06401040405035019,
+ "learning_rate": 0.00026102037482462256,
+ "loss": 1.9732205867767334,
+ "step": 2009
+ },
+ {
+ "epoch": 27.9198606271777,
+ "grad_norm": 0.06437943130731583,
+ "learning_rate": 0.0002607564816380403,
+ "loss": 1.974897027015686,
+ "step": 2010
+ },
+ {
+ "epoch": 27.933797909407666,
+ "grad_norm": 0.06296383589506149,
+ "learning_rate": 0.0002604926224862975,
+ "loss": 1.9710102081298828,
+ "step": 2011
+ },
+ {
+ "epoch": 27.94773519163763,
+ "grad_norm": 0.06527025997638702,
+ "learning_rate": 0.0002602287975819811,
+ "loss": 1.9707660675048828,
+ "step": 2012
+ },
+ {
+ "epoch": 27.961672473867594,
+ "grad_norm": 0.06185369938611984,
+ "learning_rate": 0.00025996500713765,
+ "loss": 1.9675114154815674,
+ "step": 2013
+ },
+ {
+ "epoch": 27.975609756097562,
+ "grad_norm": 0.06357269734144211,
+ "learning_rate": 0.0002597012513658352,
+ "loss": 1.9812211990356445,
+ "step": 2014
+ },
+ {
+ "epoch": 27.989547038327526,
+ "grad_norm": 0.06581240892410278,
+ "learning_rate": 0.00025943753047904026,
+ "loss": 1.97153639793396,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.049184102565050125,
+ "learning_rate": 0.0002591738446897401,
+ "loss": 1.4647765159606934,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 0.6328383088111877,
+ "eval_runtime": 40.5927,
+ "eval_samples_per_second": 60.159,
+ "eval_steps_per_second": 0.493,
+ "step": 2016
+ },
+ {
+ "epoch": 28.013937282229964,
+ "grad_norm": 0.05999046564102173,
+ "learning_rate": 0.00025891019421038177,
+ "loss": 1.9182333946228027,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027874564459932,
+ "grad_norm": 0.059804853051900864,
+ "learning_rate": 0.00025864657925338345,
+ "loss": 1.924762487411499,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041811846689896,
+ "grad_norm": 0.05890650674700737,
+ "learning_rate": 0.00025838300003113514,
+ "loss": 1.9341628551483154,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05574912891986,
+ "grad_norm": 0.057233694940805435,
+ "learning_rate": 0.0002581194567559979,
+ "loss": 1.932476282119751,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069686411149824,
+ "grad_norm": 0.060101717710494995,
+ "learning_rate": 0.00025785594964030365,
+ "loss": 1.9403120279312134,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083623693379792,
+ "grad_norm": 0.06054912135004997,
+ "learning_rate": 0.0002575924788963554,
+ "loss": 1.9274934530258179,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097560975609756,
+ "grad_norm": 0.05882031098008156,
+ "learning_rate": 0.0002573290447364268,
+ "loss": 1.9331915378570557,
+ "step": 2023
+ },
+ {
+ "epoch": 28.11149825783972,
+ "grad_norm": 0.05846688151359558,
+ "learning_rate": 0.0002570656473727619,
+ "loss": 1.9342529773712158,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125435540069688,
+ "grad_norm": 0.058974213898181915,
+ "learning_rate": 0.0002568022870175752,
+ "loss": 1.9353152513504028,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139372822299652,
+ "grad_norm": 0.05882132798433304,
+ "learning_rate": 0.00025653896388305134,
+ "loss": 1.9500072002410889,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153310104529616,
+ "grad_norm": 0.05919627472758293,
+ "learning_rate": 0.0002562756781813453,
+ "loss": 1.9313960075378418,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16724738675958,
+ "grad_norm": 0.06413958966732025,
+ "learning_rate": 0.00025601243012458126,
+ "loss": 1.9516873359680176,
+ "step": 2028
+ },
+ {
+ "epoch": 28.181184668989548,
+ "grad_norm": 0.05799189209938049,
+ "learning_rate": 0.00025574921992485366,
+ "loss": 1.9290452003479004,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195121951219512,
+ "grad_norm": 0.05999152734875679,
+ "learning_rate": 0.000255486047794226,
+ "loss": 1.9396628141403198,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209059233449477,
+ "grad_norm": 0.060240138322114944,
+ "learning_rate": 0.0002552229139447316,
+ "loss": 1.937988519668579,
+ "step": 2031
+ },
+ {
+ "epoch": 28.222996515679444,
+ "grad_norm": 0.05933740735054016,
+ "learning_rate": 0.00025495981858837235,
+ "loss": 1.9343246221542358,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23693379790941,
+ "grad_norm": 0.05960093438625336,
+ "learning_rate": 0.0002546967619371195,
+ "loss": 1.9326903820037842,
+ "step": 2033
+ },
+ {
+ "epoch": 28.250871080139373,
+ "grad_norm": 0.0635334923863411,
+ "learning_rate": 0.0002544337442029133,
+ "loss": 1.9381543397903442,
+ "step": 2034
+ },
+ {
+ "epoch": 28.264808362369337,
+ "grad_norm": 0.06162877380847931,
+ "learning_rate": 0.00025417076559766194,
+ "loss": 1.9265705347061157,
+ "step": 2035
+ },
+ {
+ "epoch": 28.278745644599304,
+ "grad_norm": 0.059690047055482864,
+ "learning_rate": 0.000253907826333243,
+ "loss": 1.9278099536895752,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29268292682927,
+ "grad_norm": 0.0612013153731823,
+ "learning_rate": 0.00025364492662150145,
+ "loss": 1.939026117324829,
+ "step": 2037
+ },
+ {
+ "epoch": 28.306620209059233,
+ "grad_norm": 0.06080419942736626,
+ "learning_rate": 0.0002533820666742514,
+ "loss": 1.9306461811065674,
+ "step": 2038
+ },
+ {
+ "epoch": 28.320557491289197,
+ "grad_norm": 0.06308622658252716,
+ "learning_rate": 0.00025311924670327364,
+ "loss": 1.9350788593292236,
+ "step": 2039
+ },
+ {
+ "epoch": 28.334494773519165,
+ "grad_norm": 0.06124545633792877,
+ "learning_rate": 0.00025285646692031804,
+ "loss": 1.9264733791351318,
+ "step": 2040
+ },
+ {
+ "epoch": 28.34843205574913,
+ "grad_norm": 0.059037938714027405,
+ "learning_rate": 0.0002525937275371013,
+ "loss": 1.917725920677185,
+ "step": 2041
+ },
+ {
+ "epoch": 28.362369337979093,
+ "grad_norm": 0.06089222431182861,
+ "learning_rate": 0.00025233102876530785,
+ "loss": 1.949327826499939,
+ "step": 2042
+ },
+ {
+ "epoch": 28.37630662020906,
+ "grad_norm": 0.05957084521651268,
+ "learning_rate": 0.0002520683708165894,
+ "loss": 1.938978910446167,
+ "step": 2043
+ },
+ {
+ "epoch": 28.390243902439025,
+ "grad_norm": 0.06060048192739487,
+ "learning_rate": 0.0002518057539025646,
+ "loss": 1.9226362705230713,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40418118466899,
+ "grad_norm": 0.058304425328969955,
+ "learning_rate": 0.00025154317823481917,
+ "loss": 1.9330644607543945,
+ "step": 2045
+ },
+ {
+ "epoch": 28.418118466898953,
+ "grad_norm": 0.05970550701022148,
+ "learning_rate": 0.0002512806440249056,
+ "loss": 1.931627631187439,
+ "step": 2046
+ },
+ {
+ "epoch": 28.43205574912892,
+ "grad_norm": 0.05994785949587822,
+ "learning_rate": 0.00025101815148434296,
+ "loss": 1.9387202262878418,
+ "step": 2047
+ },
+ {
+ "epoch": 28.445993031358885,
+ "grad_norm": 0.06001518666744232,
+ "learning_rate": 0.00025075570082461684,
+ "loss": 1.951300859451294,
+ "step": 2048
+ },
+ {
+ "epoch": 28.45993031358885,
+ "grad_norm": 0.0636107549071312,
+ "learning_rate": 0.0002504932922571789,
+ "loss": 1.9358865022659302,
+ "step": 2049
+ },
+ {
+ "epoch": 28.473867595818817,
+ "grad_norm": 0.059402115643024445,
+ "learning_rate": 0.00025023092599344706,
+ "loss": 1.9298653602600098,
+ "step": 2050
+ },
+ {
+ "epoch": 28.48780487804878,
+ "grad_norm": 0.06260179728269577,
+ "learning_rate": 0.0002499686022448049,
+ "loss": 1.954633116722107,
+ "step": 2051
+ },
+ {
+ "epoch": 28.501742160278745,
+ "grad_norm": 0.06439191848039627,
+ "learning_rate": 0.00024970632122260216,
+ "loss": 1.9338306188583374,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51567944250871,
+ "grad_norm": 0.06265310943126678,
+ "learning_rate": 0.0002494440831381538,
+ "loss": 1.9500043392181396,
+ "step": 2053
+ },
+ {
+ "epoch": 28.529616724738677,
+ "grad_norm": 0.06323659420013428,
+ "learning_rate": 0.00024918188820274016,
+ "loss": 1.9445586204528809,
+ "step": 2054
+ },
+ {
+ "epoch": 28.54355400696864,
+ "grad_norm": 0.06512445956468582,
+ "learning_rate": 0.0002489197366276073,
+ "loss": 1.958587408065796,
+ "step": 2055
+ },
+ {
+ "epoch": 28.557491289198605,
+ "grad_norm": 0.06386571377515793,
+ "learning_rate": 0.0002486576286239657,
+ "loss": 1.949887990951538,
+ "step": 2056
+ },
+ {
+ "epoch": 28.571428571428573,
+ "grad_norm": 0.06317705661058426,
+ "learning_rate": 0.0002483955644029913,
+ "loss": 1.9512674808502197,
+ "step": 2057
+ },
+ {
+ "epoch": 28.585365853658537,
+ "grad_norm": 0.06504984200000763,
+ "learning_rate": 0.0002481335441758243,
+ "loss": 1.9242960214614868,
+ "step": 2058
+ },
+ {
+ "epoch": 28.5993031358885,
+ "grad_norm": 0.06278593838214874,
+ "learning_rate": 0.0002478715681535699,
+ "loss": 1.9203131198883057,
+ "step": 2059
+ },
+ {
+ "epoch": 28.613240418118465,
+ "grad_norm": 0.06171257048845291,
+ "learning_rate": 0.00024760963654729717,
+ "loss": 1.954156517982483,
+ "step": 2060
+ },
+ {
+ "epoch": 28.627177700348433,
+ "grad_norm": 0.06303241103887558,
+ "learning_rate": 0.0002473477495680397,
+ "loss": 1.9460458755493164,
+ "step": 2061
+ },
+ {
+ "epoch": 28.641114982578397,
+ "grad_norm": 0.06252635270357132,
+ "learning_rate": 0.00024708590742679546,
+ "loss": 1.9477293491363525,
+ "step": 2062
+ },
+ {
+ "epoch": 28.65505226480836,
+ "grad_norm": 0.06226412579417229,
+ "learning_rate": 0.00024682411033452546,
+ "loss": 1.9259647130966187,
+ "step": 2063
+ },
+ {
+ "epoch": 28.66898954703833,
+ "grad_norm": 0.06111626327037811,
+ "learning_rate": 0.0002465623585021553,
+ "loss": 1.9573581218719482,
+ "step": 2064
+ },
+ {
+ "epoch": 28.682926829268293,
+ "grad_norm": 0.06460461020469666,
+ "learning_rate": 0.0002463006521405733,
+ "loss": 1.9486393928527832,
+ "step": 2065
+ },
+ {
+ "epoch": 28.696864111498257,
+ "grad_norm": 0.062027834355831146,
+ "learning_rate": 0.0002460389914606319,
+ "loss": 1.9404425621032715,
+ "step": 2066
+ },
+ {
+ "epoch": 28.71080139372822,
+ "grad_norm": 0.061942074447870255,
+ "learning_rate": 0.00024577737667314615,
+ "loss": 1.9444060325622559,
+ "step": 2067
+ },
+ {
+ "epoch": 28.72473867595819,
+ "grad_norm": 0.06159934028983116,
+ "learning_rate": 0.00024551580798889446,
+ "loss": 1.9635517597198486,
+ "step": 2068
+ },
+ {
+ "epoch": 28.738675958188153,
+ "grad_norm": 0.0621839202940464,
+ "learning_rate": 0.00024525428561861804,
+ "loss": 1.9442765712738037,
+ "step": 2069
+ },
+ {
+ "epoch": 28.752613240418118,
+ "grad_norm": 0.06432677060365677,
+ "learning_rate": 0.0002449928097730207,
+ "loss": 1.9604840278625488,
+ "step": 2070
+ },
+ {
+ "epoch": 28.766550522648085,
+ "grad_norm": 0.06037316843867302,
+ "learning_rate": 0.00024473138066276876,
+ "loss": 1.948632836341858,
+ "step": 2071
+ },
+ {
+ "epoch": 28.78048780487805,
+ "grad_norm": 0.05999939888715744,
+ "learning_rate": 0.0002444699984984909,
+ "loss": 1.9451751708984375,
+ "step": 2072
+ },
+ {
+ "epoch": 28.794425087108014,
+ "grad_norm": 0.06441698968410492,
+ "learning_rate": 0.0002442086634907782,
+ "loss": 1.9356517791748047,
+ "step": 2073
+ },
+ {
+ "epoch": 28.808362369337978,
+ "grad_norm": 0.06059470772743225,
+ "learning_rate": 0.00024394737585018328,
+ "loss": 1.9258331060409546,
+ "step": 2074
+ },
+ {
+ "epoch": 28.822299651567945,
+ "grad_norm": 0.062317874282598495,
+ "learning_rate": 0.00024368613578722113,
+ "loss": 1.9342000484466553,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83623693379791,
+ "grad_norm": 0.06318258494138718,
+ "learning_rate": 0.00024342494351236799,
+ "loss": 1.9395548105239868,
+ "step": 2076
+ },
+ {
+ "epoch": 28.850174216027874,
+ "grad_norm": 0.06204817444086075,
+ "learning_rate": 0.0002431637992360618,
+ "loss": 1.9572374820709229,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86411149825784,
+ "grad_norm": 0.06250188499689102,
+ "learning_rate": 0.00024290270316870184,
+ "loss": 1.9465739727020264,
+ "step": 2078
+ },
+ {
+ "epoch": 28.878048780487806,
+ "grad_norm": 0.06293781846761703,
+ "learning_rate": 0.0002426416555206484,
+ "loss": 1.9431169033050537,
+ "step": 2079
+ },
+ {
+ "epoch": 28.89198606271777,
+ "grad_norm": 0.059349458664655685,
+ "learning_rate": 0.000242380656502223,
+ "loss": 1.9673197269439697,
+ "step": 2080
+ },
+ {
+ "epoch": 28.905923344947734,
+ "grad_norm": 0.06214660778641701,
+ "learning_rate": 0.00024211970632370756,
+ "loss": 1.9646787643432617,
+ "step": 2081
+ },
+ {
+ "epoch": 28.9198606271777,
+ "grad_norm": 0.062279123812913895,
+ "learning_rate": 0.0002418588051953453,
+ "loss": 1.9572803974151611,
+ "step": 2082
+ },
+ {
+ "epoch": 28.933797909407666,
+ "grad_norm": 0.06502588093280792,
+ "learning_rate": 0.0002415979533273392,
+ "loss": 1.9510462284088135,
+ "step": 2083
+ },
+ {
+ "epoch": 28.94773519163763,
+ "grad_norm": 0.06153101101517677,
+ "learning_rate": 0.0002413371509298531,
+ "loss": 1.9507923126220703,
+ "step": 2084
+ },
+ {
+ "epoch": 28.961672473867594,
+ "grad_norm": 0.06469229608774185,
+ "learning_rate": 0.00024107639821301078,
+ "loss": 1.937751293182373,
+ "step": 2085
+ },
+ {
+ "epoch": 28.975609756097562,
+ "grad_norm": 0.0672568678855896,
+ "learning_rate": 0.00024081569538689587,
+ "loss": 1.944124698638916,
+ "step": 2086
+ },
+ {
+ "epoch": 28.989547038327526,
+ "grad_norm": 0.06451335549354553,
+ "learning_rate": 0.0002405550426615521,
+ "loss": 1.9519702196121216,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.05141456052660942,
+ "learning_rate": 0.00024029444024698248,
+ "loss": 1.4445865154266357,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 0.6343450546264648,
+ "eval_runtime": 40.5081,
+ "eval_samples_per_second": 60.284,
+ "eval_steps_per_second": 0.494,
+ "step": 2088
+ },
+ {
+ "epoch": 29.013937282229964,
+ "grad_norm": 0.061562713235616684,
+ "learning_rate": 0.00024003388835314981,
+ "loss": 1.921946406364441,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027874564459932,
+ "grad_norm": 0.06396839767694473,
+ "learning_rate": 0.00023977338718997595,
+ "loss": 1.9222416877746582,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041811846689896,
+ "grad_norm": 0.05828692391514778,
+ "learning_rate": 0.00023951293696734195,
+ "loss": 1.9173905849456787,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05574912891986,
+ "grad_norm": 0.06567556411027908,
+ "learning_rate": 0.00023925253789508806,
+ "loss": 1.9148579835891724,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069686411149824,
+ "grad_norm": 0.05924173817038536,
+ "learning_rate": 0.00023899219018301286,
+ "loss": 1.916333794593811,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083623693379792,
+ "grad_norm": 0.06129243224859238,
+ "learning_rate": 0.00023873189404087405,
+ "loss": 1.9031038284301758,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097560975609756,
+ "grad_norm": 0.06068028509616852,
+ "learning_rate": 0.00023847164967838726,
+ "loss": 1.9165983200073242,
+ "step": 2095
+ },
+ {
+ "epoch": 29.11149825783972,
+ "grad_norm": 0.06304074823856354,
+ "learning_rate": 0.00023821145730522692,
+ "loss": 1.90432870388031,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125435540069688,
+ "grad_norm": 0.060179345309734344,
+ "learning_rate": 0.0002379513171310251,
+ "loss": 1.9187151193618774,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139372822299652,
+ "grad_norm": 0.05633116513490677,
+ "learning_rate": 0.00023769122936537207,
+ "loss": 1.9125087261199951,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153310104529616,
+ "grad_norm": 0.0598532073199749,
+ "learning_rate": 0.00023743119421781614,
+ "loss": 1.906031847000122,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16724738675958,
+ "grad_norm": 0.05987119674682617,
+ "learning_rate": 0.00023717121189786256,
+ "loss": 1.9177607297897339,
+ "step": 2100
+ },
+ {
+ "epoch": 29.181184668989548,
+ "grad_norm": 0.060883522033691406,
+ "learning_rate": 0.00023691128261497463,
+ "loss": 1.9008433818817139,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195121951219512,
+ "grad_norm": 0.05977628752589226,
+ "learning_rate": 0.0002366514065785725,
+ "loss": 1.9176266193389893,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209059233449477,
+ "grad_norm": 0.05939047038555145,
+ "learning_rate": 0.00023639158399803382,
+ "loss": 1.9227800369262695,
+ "step": 2103
+ },
+ {
+ "epoch": 29.222996515679444,
+ "grad_norm": 0.059280600398778915,
+ "learning_rate": 0.0002361318150826927,
+ "loss": 1.9164855480194092,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23693379790941,
+ "grad_norm": 0.058746229857206345,
+ "learning_rate": 0.0002358721000418404,
+ "loss": 1.9110995531082153,
+ "step": 2105
+ },
+ {
+ "epoch": 29.250871080139373,
+ "grad_norm": 0.05882417410612106,
+ "learning_rate": 0.00023561243908472472,
+ "loss": 1.926957368850708,
+ "step": 2106
+ },
+ {
+ "epoch": 29.264808362369337,
+ "grad_norm": 0.05827188491821289,
+ "learning_rate": 0.00023535283242054954,
+ "loss": 1.9101874828338623,
+ "step": 2107
+ },
+ {
+ "epoch": 29.278745644599304,
+ "grad_norm": 0.059739116579294205,
+ "learning_rate": 0.00023509328025847565,
+ "loss": 1.924944519996643,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29268292682927,
+ "grad_norm": 0.058013755828142166,
+ "learning_rate": 0.00023483378280761907,
+ "loss": 1.9121344089508057,
+ "step": 2109
+ },
+ {
+ "epoch": 29.306620209059233,
+ "grad_norm": 0.061207350343465805,
+ "learning_rate": 0.0002345743402770527,
+ "loss": 1.9137616157531738,
+ "step": 2110
+ },
+ {
+ "epoch": 29.320557491289197,
+ "grad_norm": 0.05848725140094757,
+ "learning_rate": 0.0002343149528758042,
+ "loss": 1.9290939569473267,
+ "step": 2111
+ },
+ {
+ "epoch": 29.334494773519165,
+ "grad_norm": 0.059134867042303085,
+ "learning_rate": 0.00023405562081285766,
+ "loss": 1.919262409210205,
+ "step": 2112
+ },
+ {
+ "epoch": 29.34843205574913,
+ "grad_norm": 0.06025470420718193,
+ "learning_rate": 0.0002337963442971522,
+ "loss": 1.9021251201629639,
+ "step": 2113
+ },
+ {
+ "epoch": 29.362369337979093,
+ "grad_norm": 0.05981536582112312,
+ "learning_rate": 0.00023353712353758212,
+ "loss": 1.9163634777069092,
+ "step": 2114
+ },
+ {
+ "epoch": 29.37630662020906,
+ "grad_norm": 0.05826934054493904,
+ "learning_rate": 0.00023327795874299704,
+ "loss": 1.8987491130828857,
+ "step": 2115
+ },
+ {
+ "epoch": 29.390243902439025,
+ "grad_norm": 0.06095297262072563,
+ "learning_rate": 0.0002330188501222012,
+ "loss": 1.9322285652160645,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40418118466899,
+ "grad_norm": 0.05814153701066971,
+ "learning_rate": 0.0002327597978839538,
+ "loss": 1.927016019821167,
+ "step": 2117
+ },
+ {
+ "epoch": 29.418118466898953,
+ "grad_norm": 0.05872226879000664,
+ "learning_rate": 0.00023250080223696852,
+ "loss": 1.9271752834320068,
+ "step": 2118
+ },
+ {
+ "epoch": 29.43205574912892,
+ "grad_norm": 0.06008293107151985,
+ "learning_rate": 0.00023224186338991352,
+ "loss": 1.9269304275512695,
+ "step": 2119
+ },
+ {
+ "epoch": 29.445993031358885,
+ "grad_norm": 0.06035081297159195,
+ "learning_rate": 0.0002319829815514112,
+ "loss": 1.929135799407959,
+ "step": 2120
+ },
+ {
+ "epoch": 29.45993031358885,
+ "grad_norm": 0.06080561876296997,
+ "learning_rate": 0.00023172415693003789,
+ "loss": 1.9284312725067139,
+ "step": 2121
+ },
+ {
+ "epoch": 29.473867595818817,
+ "grad_norm": 0.058793697506189346,
+ "learning_rate": 0.000231465389734324,
+ "loss": 1.931719183921814,
+ "step": 2122
+ },
+ {
+ "epoch": 29.48780487804878,
+ "grad_norm": 0.06385456770658493,
+ "learning_rate": 0.00023120668017275346,
+ "loss": 1.9112051725387573,
+ "step": 2123
+ },
+ {
+ "epoch": 29.501742160278745,
+ "grad_norm": 0.06030426174402237,
+ "learning_rate": 0.00023094802845376403,
+ "loss": 1.9254870414733887,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51567944250871,
+ "grad_norm": 0.059933315962553024,
+ "learning_rate": 0.00023068943478574662,
+ "loss": 1.915220022201538,
+ "step": 2125
+ },
+ {
+ "epoch": 29.529616724738677,
+ "grad_norm": 0.06083494424819946,
+ "learning_rate": 0.00023043089937704541,
+ "loss": 1.9391965866088867,
+ "step": 2126
+ },
+ {
+ "epoch": 29.54355400696864,
+ "grad_norm": 0.06049460917711258,
+ "learning_rate": 0.00023017242243595796,
+ "loss": 1.9301774501800537,
+ "step": 2127
+ },
+ {
+ "epoch": 29.557491289198605,
+ "grad_norm": 0.05953316017985344,
+ "learning_rate": 0.00022991400417073427,
+ "loss": 1.9150071144104004,
+ "step": 2128
+ },
+ {
+ "epoch": 29.571428571428573,
+ "grad_norm": 0.0611971914768219,
+ "learning_rate": 0.00022965564478957739,
+ "loss": 1.9265213012695312,
+ "step": 2129
+ },
+ {
+ "epoch": 29.585365853658537,
+ "grad_norm": 0.06123765930533409,
+ "learning_rate": 0.0002293973445006427,
+ "loss": 1.9405107498168945,
+ "step": 2130
+ },
+ {
+ "epoch": 29.5993031358885,
+ "grad_norm": 0.06403640657663345,
+ "learning_rate": 0.00022913910351203816,
+ "loss": 1.9216606616973877,
+ "step": 2131
+ },
+ {
+ "epoch": 29.613240418118465,
+ "grad_norm": 0.06530711054801941,
+ "learning_rate": 0.00022888092203182377,
+ "loss": 1.9302153587341309,
+ "step": 2132
+ },
+ {
+ "epoch": 29.627177700348433,
+ "grad_norm": 0.06106860563158989,
+ "learning_rate": 0.00022862280026801182,
+ "loss": 1.930979609489441,
+ "step": 2133
+ },
+ {
+ "epoch": 29.641114982578397,
+ "grad_norm": 0.06148209050297737,
+ "learning_rate": 0.00022836473842856623,
+ "loss": 1.9276363849639893,
+ "step": 2134
+ },
+ {
+ "epoch": 29.65505226480836,
+ "grad_norm": 0.060283321887254715,
+ "learning_rate": 0.00022810673672140272,
+ "loss": 1.9269459247589111,
+ "step": 2135
+ },
+ {
+ "epoch": 29.66898954703833,
+ "grad_norm": 0.06138703227043152,
+ "learning_rate": 0.0002278487953543889,
+ "loss": 1.9389945268630981,
+ "step": 2136
+ },
+ {
+ "epoch": 29.682926829268293,
+ "grad_norm": 0.06101245433092117,
+ "learning_rate": 0.00022759091453534308,
+ "loss": 1.9426372051239014,
+ "step": 2137
+ },
+ {
+ "epoch": 29.696864111498257,
+ "grad_norm": 0.05958810821175575,
+ "learning_rate": 0.00022733309447203553,
+ "loss": 1.9327162504196167,
+ "step": 2138
+ },
+ {
+ "epoch": 29.71080139372822,
+ "grad_norm": 0.06250876933336258,
+ "learning_rate": 0.00022707533537218695,
+ "loss": 1.9384305477142334,
+ "step": 2139
+ },
+ {
+ "epoch": 29.72473867595819,
+ "grad_norm": 0.06344381719827652,
+ "learning_rate": 0.00022681763744346938,
+ "loss": 1.9422640800476074,
+ "step": 2140
+ },
+ {
+ "epoch": 29.738675958188153,
+ "grad_norm": 0.06065250560641289,
+ "learning_rate": 0.00022656000089350524,
+ "loss": 1.920068621635437,
+ "step": 2141
+ },
+ {
+ "epoch": 29.752613240418118,
+ "grad_norm": 0.06333960592746735,
+ "learning_rate": 0.00022630242592986766,
+ "loss": 1.925317645072937,
+ "step": 2142
+ },
+ {
+ "epoch": 29.766550522648085,
+ "grad_norm": 0.06285370141267776,
+ "learning_rate": 0.0002260449127600803,
+ "loss": 1.92912757396698,
+ "step": 2143
+ },
+ {
+ "epoch": 29.78048780487805,
+ "grad_norm": 0.06441807746887207,
+ "learning_rate": 0.0002257874615916166,
+ "loss": 1.941272497177124,
+ "step": 2144
+ },
+ {
+ "epoch": 29.794425087108014,
+ "grad_norm": 0.06314355880022049,
+ "learning_rate": 0.00022553007263190067,
+ "loss": 1.923574447631836,
+ "step": 2145
+ },
+ {
+ "epoch": 29.808362369337978,
+ "grad_norm": 0.0639389157295227,
+ "learning_rate": 0.00022527274608830574,
+ "loss": 1.9339193105697632,
+ "step": 2146
+ },
+ {
+ "epoch": 29.822299651567945,
+ "grad_norm": 0.06419187039136887,
+ "learning_rate": 0.0002250154821681555,
+ "loss": 1.9294497966766357,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83623693379791,
+ "grad_norm": 0.06351246684789658,
+ "learning_rate": 0.0002247582810787227,
+ "loss": 1.9282407760620117,
+ "step": 2148
+ },
+ {
+ "epoch": 29.850174216027874,
+ "grad_norm": 0.06371650099754333,
+ "learning_rate": 0.00022450114302722957,
+ "loss": 1.914009690284729,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86411149825784,
+ "grad_norm": 0.06891132146120071,
+ "learning_rate": 0.0002242440682208477,
+ "loss": 1.948460578918457,
+ "step": 2150
+ },
+ {
+ "epoch": 29.878048780487806,
+ "grad_norm": 0.06575452536344528,
+ "learning_rate": 0.00022398705686669748,
+ "loss": 1.9238896369934082,
+ "step": 2151
+ },
+ {
+ "epoch": 29.89198606271777,
+ "grad_norm": 0.06279128044843674,
+ "learning_rate": 0.0002237301091718485,
+ "loss": 1.9414188861846924,
+ "step": 2152
+ },
+ {
+ "epoch": 29.905923344947734,
+ "grad_norm": 0.06493876129388809,
+ "learning_rate": 0.00022347322534331865,
+ "loss": 1.9337334632873535,
+ "step": 2153
+ },
+ {
+ "epoch": 29.9198606271777,
+ "grad_norm": 0.0640307068824768,
+ "learning_rate": 0.00022321640558807488,
+ "loss": 1.9258478879928589,
+ "step": 2154
+ },
+ {
+ "epoch": 29.933797909407666,
+ "grad_norm": 0.06517527252435684,
+ "learning_rate": 0.00022295965011303188,
+ "loss": 1.9236748218536377,
+ "step": 2155
+ },
+ {
+ "epoch": 29.94773519163763,
+ "grad_norm": 0.06373083591461182,
+ "learning_rate": 0.00022270295912505317,
+ "loss": 1.9360520839691162,
+ "step": 2156
+ },
+ {
+ "epoch": 29.961672473867594,
+ "grad_norm": 0.06771238893270493,
+ "learning_rate": 0.00022244633283095,
+ "loss": 1.947780966758728,
+ "step": 2157
+ },
+ {
+ "epoch": 29.975609756097562,
+ "grad_norm": 0.06298559159040451,
+ "learning_rate": 0.0002221897714374814,
+ "loss": 1.9223754405975342,
+ "step": 2158
+ },
+ {
+ "epoch": 29.989547038327526,
+ "grad_norm": 0.06455206871032715,
+ "learning_rate": 0.00022193327515135445,
+ "loss": 1.9330079555511475,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.0522056445479393,
+ "learning_rate": 0.00022167684417922337,
+ "loss": 1.4463932514190674,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 0.6358165740966797,
+ "eval_runtime": 40.2782,
+ "eval_samples_per_second": 60.628,
+ "eval_steps_per_second": 0.497,
+ "step": 2160
+ },
+ {
+ "epoch": 30.013937282229964,
+ "grad_norm": 0.06026934087276459,
+ "learning_rate": 0.00022142047872769014,
+ "loss": 1.9035022258758545,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027874564459932,
+ "grad_norm": 0.06600011885166168,
+ "learning_rate": 0.00022116417900330365,
+ "loss": 1.8965950012207031,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041811846689896,
+ "grad_norm": 0.060655586421489716,
+ "learning_rate": 0.00022090794521255997,
+ "loss": 1.8954249620437622,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05574912891986,
+ "grad_norm": 0.0636422410607338,
+ "learning_rate": 0.00022065177756190214,
+ "loss": 1.8951303958892822,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069686411149824,
+ "grad_norm": 0.05972317233681679,
+ "learning_rate": 0.00022039567625771968,
+ "loss": 1.910957932472229,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083623693379792,
+ "grad_norm": 0.060592107474803925,
+ "learning_rate": 0.0002201396415063489,
+ "loss": 1.9128470420837402,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097560975609756,
+ "grad_norm": 0.05888872966170311,
+ "learning_rate": 0.00021988367351407227,
+ "loss": 1.897517204284668,
+ "step": 2167
+ },
+ {
+ "epoch": 30.11149825783972,
+ "grad_norm": 0.062287602573633194,
+ "learning_rate": 0.00021962777248711872,
+ "loss": 1.9048705101013184,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125435540069688,
+ "grad_norm": 0.05919402465224266,
+ "learning_rate": 0.00021937193863166292,
+ "loss": 1.901914119720459,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139372822299652,
+ "grad_norm": 0.06069913133978844,
+ "learning_rate": 0.00021911617215382564,
+ "loss": 1.9066851139068604,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153310104529616,
+ "grad_norm": 0.06046717241406441,
+ "learning_rate": 0.00021886047325967348,
+ "loss": 1.8969058990478516,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16724738675958,
+ "grad_norm": 0.059624481946229935,
+ "learning_rate": 0.00021860484215521805,
+ "loss": 1.8908498287200928,
+ "step": 2172
+ },
+ {
+ "epoch": 30.181184668989548,
+ "grad_norm": 0.06100156158208847,
+ "learning_rate": 0.00021834927904641708,
+ "loss": 1.909839391708374,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195121951219512,
+ "grad_norm": 0.05801135674118996,
+ "learning_rate": 0.00021809378413917296,
+ "loss": 1.9057464599609375,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209059233449477,
+ "grad_norm": 0.06198010593652725,
+ "learning_rate": 0.00021783835763933344,
+ "loss": 1.9051663875579834,
+ "step": 2175
+ },
+ {
+ "epoch": 30.222996515679444,
+ "grad_norm": 0.06107710674405098,
+ "learning_rate": 0.0002175829997526909,
+ "loss": 1.904172658920288,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23693379790941,
+ "grad_norm": 0.060085635632276535,
+ "learning_rate": 0.0002173277106849826,
+ "loss": 1.8970673084259033,
+ "step": 2177
+ },
+ {
+ "epoch": 30.250871080139373,
+ "grad_norm": 0.05755467712879181,
+ "learning_rate": 0.0002170724906418905,
+ "loss": 1.9092798233032227,
+ "step": 2178
+ },
+ {
+ "epoch": 30.264808362369337,
+ "grad_norm": 0.05827959626913071,
+ "learning_rate": 0.00021681733982904046,
+ "loss": 1.9012596607208252,
+ "step": 2179
+ },
+ {
+ "epoch": 30.278745644599304,
+ "grad_norm": 0.0570731982588768,
+ "learning_rate": 0.00021656225845200332,
+ "loss": 1.9027369022369385,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29268292682927,
+ "grad_norm": 0.05898377671837807,
+ "learning_rate": 0.0002163072467162931,
+ "loss": 1.9284251928329468,
+ "step": 2181
+ },
+ {
+ "epoch": 30.306620209059233,
+ "grad_norm": 0.06119367107748985,
+ "learning_rate": 0.00021605230482736845,
+ "loss": 1.8992201089859009,
+ "step": 2182
+ },
+ {
+ "epoch": 30.320557491289197,
+ "grad_norm": 0.05790785327553749,
+ "learning_rate": 0.00021579743299063122,
+ "loss": 1.9106863737106323,
+ "step": 2183
+ },
+ {
+ "epoch": 30.334494773519165,
+ "grad_norm": 0.06101157143712044,
+ "learning_rate": 0.00021554263141142726,
+ "loss": 1.9129860401153564,
+ "step": 2184
+ },
+ {
+ "epoch": 30.34843205574913,
+ "grad_norm": 0.062228068709373474,
+ "learning_rate": 0.00021528790029504548,
+ "loss": 1.9075040817260742,
+ "step": 2185
+ },
+ {
+ "epoch": 30.362369337979093,
+ "grad_norm": 0.057724758982658386,
+ "learning_rate": 0.00021503323984671814,
+ "loss": 1.9021648168563843,
+ "step": 2186
+ },
+ {
+ "epoch": 30.37630662020906,
+ "grad_norm": 0.05886731296777725,
+ "learning_rate": 0.00021477865027162067,
+ "loss": 1.8955028057098389,
+ "step": 2187
+ },
+ {
+ "epoch": 30.390243902439025,
+ "grad_norm": 0.0613555908203125,
+ "learning_rate": 0.0002145241317748711,
+ "loss": 1.8902294635772705,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40418118466899,
+ "grad_norm": 0.059424031525850296,
+ "learning_rate": 0.0002142696845615306,
+ "loss": 1.9234192371368408,
+ "step": 2189
+ },
+ {
+ "epoch": 30.418118466898953,
+ "grad_norm": 0.059314578771591187,
+ "learning_rate": 0.0002140153088366024,
+ "loss": 1.9120237827301025,
+ "step": 2190
+ },
+ {
+ "epoch": 30.43205574912892,
+ "grad_norm": 0.059916652739048004,
+ "learning_rate": 0.00021376100480503276,
+ "loss": 1.9118289947509766,
+ "step": 2191
+ },
+ {
+ "epoch": 30.445993031358885,
+ "grad_norm": 0.05817565694451332,
+ "learning_rate": 0.00021350677267170973,
+ "loss": 1.9012054204940796,
+ "step": 2192
+ },
+ {
+ "epoch": 30.45993031358885,
+ "grad_norm": 0.061957065016031265,
+ "learning_rate": 0.00021325261264146342,
+ "loss": 1.9110759496688843,
+ "step": 2193
+ },
+ {
+ "epoch": 30.473867595818817,
+ "grad_norm": 0.05899856239557266,
+ "learning_rate": 0.00021299852491906616,
+ "loss": 1.9285509586334229,
+ "step": 2194
+ },
+ {
+ "epoch": 30.48780487804878,
+ "grad_norm": 0.06001519039273262,
+ "learning_rate": 0.00021274450970923167,
+ "loss": 1.9093801975250244,
+ "step": 2195
+ },
+ {
+ "epoch": 30.501742160278745,
+ "grad_norm": 0.060825154185295105,
+ "learning_rate": 0.0002124905672166156,
+ "loss": 1.892765760421753,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51567944250871,
+ "grad_norm": 0.06136799231171608,
+ "learning_rate": 0.0002122366976458146,
+ "loss": 1.9276010990142822,
+ "step": 2197
+ },
+ {
+ "epoch": 30.529616724738677,
+ "grad_norm": 0.05773079767823219,
+ "learning_rate": 0.00021198290120136702,
+ "loss": 1.9083237648010254,
+ "step": 2198
+ },
+ {
+ "epoch": 30.54355400696864,
+ "grad_norm": 0.06120939180254936,
+ "learning_rate": 0.0002117291780877519,
+ "loss": 1.8891932964324951,
+ "step": 2199
+ },
+ {
+ "epoch": 30.557491289198605,
+ "grad_norm": 0.06097486987709999,
+ "learning_rate": 0.00021147552850938941,
+ "loss": 1.9084399938583374,
+ "step": 2200
+ },
+ {
+ "epoch": 30.571428571428573,
+ "grad_norm": 0.05973029136657715,
+ "learning_rate": 0.0002112219526706406,
+ "loss": 1.918110966682434,
+ "step": 2201
+ },
+ {
+ "epoch": 30.585365853658537,
+ "grad_norm": 0.06335631012916565,
+ "learning_rate": 0.00021096845077580675,
+ "loss": 1.9174532890319824,
+ "step": 2202
+ },
+ {
+ "epoch": 30.5993031358885,
+ "grad_norm": 0.05974910035729408,
+ "learning_rate": 0.0002107150230291299,
+ "loss": 1.9107253551483154,
+ "step": 2203
+ },
+ {
+ "epoch": 30.613240418118465,
+ "grad_norm": 0.06281539052724838,
+ "learning_rate": 0.00021046166963479204,
+ "loss": 1.9007208347320557,
+ "step": 2204
+ },
+ {
+ "epoch": 30.627177700348433,
+ "grad_norm": 0.062120143324136734,
+ "learning_rate": 0.00021020839079691553,
+ "loss": 1.911742925643921,
+ "step": 2205
+ },
+ {
+ "epoch": 30.641114982578397,
+ "grad_norm": 0.06305570155382156,
+ "learning_rate": 0.00020995518671956256,
+ "loss": 1.9121983051300049,
+ "step": 2206
+ },
+ {
+ "epoch": 30.65505226480836,
+ "grad_norm": 0.06112781539559364,
+ "learning_rate": 0.00020970205760673493,
+ "loss": 1.908568263053894,
+ "step": 2207
+ },
+ {
+ "epoch": 30.66898954703833,
+ "grad_norm": 0.06050721928477287,
+ "learning_rate": 0.00020944900366237453,
+ "loss": 1.9018139839172363,
+ "step": 2208
+ },
+ {
+ "epoch": 30.682926829268293,
+ "grad_norm": 0.05980826914310455,
+ "learning_rate": 0.00020919602509036195,
+ "loss": 1.9104691743850708,
+ "step": 2209
+ },
+ {
+ "epoch": 30.696864111498257,
+ "grad_norm": 0.06286003440618515,
+ "learning_rate": 0.0002089431220945178,
+ "loss": 1.899999737739563,
+ "step": 2210
+ },
+ {
+ "epoch": 30.71080139372822,
+ "grad_norm": 0.06179053336381912,
+ "learning_rate": 0.00020869029487860126,
+ "loss": 1.8993868827819824,
+ "step": 2211
+ },
+ {
+ "epoch": 30.72473867595819,
+ "grad_norm": 0.06006763502955437,
+ "learning_rate": 0.00020843754364631068,
+ "loss": 1.9067524671554565,
+ "step": 2212
+ },
+ {
+ "epoch": 30.738675958188153,
+ "grad_norm": 0.06104302033782005,
+ "learning_rate": 0.0002081848686012832,
+ "loss": 1.9074416160583496,
+ "step": 2213
+ },
+ {
+ "epoch": 30.752613240418118,
+ "grad_norm": 0.059952687472105026,
+ "learning_rate": 0.00020793226994709446,
+ "loss": 1.9049818515777588,
+ "step": 2214
+ },
+ {
+ "epoch": 30.766550522648085,
+ "grad_norm": 0.06469111144542694,
+ "learning_rate": 0.0002076797478872588,
+ "loss": 1.9107921123504639,
+ "step": 2215
+ },
+ {
+ "epoch": 30.78048780487805,
+ "grad_norm": 0.06027381867170334,
+ "learning_rate": 0.00020742730262522844,
+ "loss": 1.920630693435669,
+ "step": 2216
+ },
+ {
+ "epoch": 30.794425087108014,
+ "grad_norm": 0.06066177412867546,
+ "learning_rate": 0.00020717493436439433,
+ "loss": 1.9131195545196533,
+ "step": 2217
+ },
+ {
+ "epoch": 30.808362369337978,
+ "grad_norm": 0.06533747911453247,
+ "learning_rate": 0.00020692264330808455,
+ "loss": 1.9172463417053223,
+ "step": 2218
+ },
+ {
+ "epoch": 30.822299651567945,
+ "grad_norm": 0.061867427080869675,
+ "learning_rate": 0.00020667042965956587,
+ "loss": 1.9100360870361328,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83623693379791,
+ "grad_norm": 0.06394216418266296,
+ "learning_rate": 0.00020641829362204195,
+ "loss": 1.9218776226043701,
+ "step": 2220
+ },
+ {
+ "epoch": 30.850174216027874,
+ "grad_norm": 0.061471469700336456,
+ "learning_rate": 0.00020616623539865446,
+ "loss": 1.901368498802185,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86411149825784,
+ "grad_norm": 0.06380587071180344,
+ "learning_rate": 0.00020591425519248213,
+ "loss": 1.9187582731246948,
+ "step": 2222
+ },
+ {
+ "epoch": 30.878048780487806,
+ "grad_norm": 0.06244281679391861,
+ "learning_rate": 0.00020566235320654069,
+ "loss": 1.9082577228546143,
+ "step": 2223
+ },
+ {
+ "epoch": 30.89198606271777,
+ "grad_norm": 0.06435921788215637,
+ "learning_rate": 0.00020541052964378323,
+ "loss": 1.924039363861084,
+ "step": 2224
+ },
+ {
+ "epoch": 30.905923344947734,
+ "grad_norm": 0.06331372261047363,
+ "learning_rate": 0.00020515878470709928,
+ "loss": 1.9083380699157715,
+ "step": 2225
+ },
+ {
+ "epoch": 30.9198606271777,
+ "grad_norm": 0.06393173336982727,
+ "learning_rate": 0.00020490711859931545,
+ "loss": 1.9132580757141113,
+ "step": 2226
+ },
+ {
+ "epoch": 30.933797909407666,
+ "grad_norm": 0.06500784307718277,
+ "learning_rate": 0.00020465553152319418,
+ "loss": 1.9109617471694946,
+ "step": 2227
+ },
+ {
+ "epoch": 30.94773519163763,
+ "grad_norm": 0.06389841437339783,
+ "learning_rate": 0.00020440402368143492,
+ "loss": 1.926948070526123,
+ "step": 2228
+ },
+ {
+ "epoch": 30.961672473867594,
+ "grad_norm": 0.06343282759189606,
+ "learning_rate": 0.00020415259527667299,
+ "loss": 1.9304101467132568,
+ "step": 2229
+ },
+ {
+ "epoch": 30.975609756097562,
+ "grad_norm": 0.06003900617361069,
+ "learning_rate": 0.0002039012465114796,
+ "loss": 1.9066693782806396,
+ "step": 2230
+ },
+ {
+ "epoch": 30.989547038327526,
+ "grad_norm": 0.06355264037847519,
+ "learning_rate": 0.00020364997758836193,
+ "loss": 1.9183707237243652,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.0492861270904541,
+ "learning_rate": 0.00020339878870976287,
+ "loss": 1.4322078227996826,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 0.6370227932929993,
+ "eval_runtime": 40.5546,
+ "eval_samples_per_second": 60.215,
+ "eval_steps_per_second": 0.493,
+ "step": 2232
+ },
+ {
+ "epoch": 31.013937282229964,
+ "grad_norm": 0.057478148490190506,
+ "learning_rate": 0.00020314768007806077,
+ "loss": 1.887700080871582,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027874564459932,
+ "grad_norm": 0.06220356002449989,
+ "learning_rate": 0.00020289665189556923,
+ "loss": 1.8875237703323364,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041811846689896,
+ "grad_norm": 0.05686607584357262,
+ "learning_rate": 0.00020264570436453707,
+ "loss": 1.897495150566101,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05574912891986,
+ "grad_norm": 0.06147320941090584,
+ "learning_rate": 0.00020239483768714843,
+ "loss": 1.8913912773132324,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069686411149824,
+ "grad_norm": 0.06075664609670639,
+ "learning_rate": 0.0002021440520655219,
+ "loss": 1.8823286294937134,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083623693379792,
+ "grad_norm": 0.058347299695014954,
+ "learning_rate": 0.00020189334770171098,
+ "loss": 1.8890151977539062,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097560975609756,
+ "grad_norm": 0.06388939172029495,
+ "learning_rate": 0.00020164272479770352,
+ "loss": 1.8857264518737793,
+ "step": 2239
+ },
+ {
+ "epoch": 31.11149825783972,
+ "grad_norm": 0.05901556834578514,
+ "learning_rate": 0.00020139218355542203,
+ "loss": 1.8779339790344238,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125435540069688,
+ "grad_norm": 0.06094517558813095,
+ "learning_rate": 0.000201141724176723,
+ "loss": 1.8811697959899902,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139372822299652,
+ "grad_norm": 0.060666318982839584,
+ "learning_rate": 0.0002008913468633969,
+ "loss": 1.8911728858947754,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153310104529616,
+ "grad_norm": 0.05608760938048363,
+ "learning_rate": 0.00020064105181716847,
+ "loss": 1.8804845809936523,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16724738675958,
+ "grad_norm": 0.06151990965008736,
+ "learning_rate": 0.00020039083923969564,
+ "loss": 1.8901598453521729,
+ "step": 2244
+ },
+ {
+ "epoch": 31.181184668989548,
+ "grad_norm": 0.05805948004126549,
+ "learning_rate": 0.0002001407093325704,
+ "loss": 1.8834660053253174,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195121951219512,
+ "grad_norm": 0.05967888608574867,
+ "learning_rate": 0.0001998906622973177,
+ "loss": 1.8807945251464844,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209059233449477,
+ "grad_norm": 0.06295159459114075,
+ "learning_rate": 0.0001996406983353961,
+ "loss": 1.8946118354797363,
+ "step": 2247
+ },
+ {
+ "epoch": 31.222996515679444,
+ "grad_norm": 0.057971853762865067,
+ "learning_rate": 0.0001993908176481969,
+ "loss": 1.889994502067566,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23693379790941,
+ "grad_norm": 0.05941799655556679,
+ "learning_rate": 0.0001991410204370445,
+ "loss": 1.884669542312622,
+ "step": 2249
+ },
+ {
+ "epoch": 31.250871080139373,
+ "grad_norm": 0.06173485890030861,
+ "learning_rate": 0.00019889130690319608,
+ "loss": 1.8727672100067139,
+ "step": 2250
+ },
+ {
+ "epoch": 31.264808362369337,
+ "grad_norm": 0.05898866057395935,
+ "learning_rate": 0.00019864167724784117,
+ "loss": 1.8795123100280762,
+ "step": 2251
+ },
+ {
+ "epoch": 31.278745644599304,
+ "grad_norm": 0.06072458624839783,
+ "learning_rate": 0.00019839213167210213,
+ "loss": 1.8846805095672607,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29268292682927,
+ "grad_norm": 0.060174115002155304,
+ "learning_rate": 0.00019814267037703294,
+ "loss": 1.8993717432022095,
+ "step": 2253
+ },
+ {
+ "epoch": 31.306620209059233,
+ "grad_norm": 0.061605289578437805,
+ "learning_rate": 0.0001978932935636205,
+ "loss": 1.8877589702606201,
+ "step": 2254
+ },
+ {
+ "epoch": 31.320557491289197,
+ "grad_norm": 0.06246590241789818,
+ "learning_rate": 0.00019764400143278276,
+ "loss": 1.8974242210388184,
+ "step": 2255
+ },
+ {
+ "epoch": 31.334494773519165,
+ "grad_norm": 0.060071494430303574,
+ "learning_rate": 0.00019739479418537018,
+ "loss": 1.881445288658142,
+ "step": 2256
+ },
+ {
+ "epoch": 31.34843205574913,
+ "grad_norm": 0.060360442847013474,
+ "learning_rate": 0.00019714567202216453,
+ "loss": 1.8938015699386597,
+ "step": 2257
+ },
+ {
+ "epoch": 31.362369337979093,
+ "grad_norm": 0.06192562356591225,
+ "learning_rate": 0.00019689663514387896,
+ "loss": 1.8961126804351807,
+ "step": 2258
+ },
+ {
+ "epoch": 31.37630662020906,
+ "grad_norm": 0.05872832611203194,
+ "learning_rate": 0.00019664768375115798,
+ "loss": 1.8974393606185913,
+ "step": 2259
+ },
+ {
+ "epoch": 31.390243902439025,
+ "grad_norm": 0.05980943515896797,
+ "learning_rate": 0.00019639881804457723,
+ "loss": 1.8886656761169434,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40418118466899,
+ "grad_norm": 0.0593055821955204,
+ "learning_rate": 0.00019615003822464342,
+ "loss": 1.8888055086135864,
+ "step": 2261
+ },
+ {
+ "epoch": 31.418118466898953,
+ "grad_norm": 0.059456177055835724,
+ "learning_rate": 0.00019590134449179371,
+ "loss": 1.9045109748840332,
+ "step": 2262
+ },
+ {
+ "epoch": 31.43205574912892,
+ "grad_norm": 0.06260113418102264,
+ "learning_rate": 0.00019565273704639648,
+ "loss": 1.9062271118164062,
+ "step": 2263
+ },
+ {
+ "epoch": 31.445993031358885,
+ "grad_norm": 0.05866454914212227,
+ "learning_rate": 0.00019540421608874994,
+ "loss": 1.8920798301696777,
+ "step": 2264
+ },
+ {
+ "epoch": 31.45993031358885,
+ "grad_norm": 0.060281749814748764,
+ "learning_rate": 0.0001951557818190831,
+ "loss": 1.9048649072647095,
+ "step": 2265
+ },
+ {
+ "epoch": 31.473867595818817,
+ "grad_norm": 0.060433730483055115,
+ "learning_rate": 0.00019490743443755504,
+ "loss": 1.8905510902404785,
+ "step": 2266
+ },
+ {
+ "epoch": 31.48780487804878,
+ "grad_norm": 0.06074376404285431,
+ "learning_rate": 0.0001946591741442546,
+ "loss": 1.890744686126709,
+ "step": 2267
+ },
+ {
+ "epoch": 31.501742160278745,
+ "grad_norm": 0.060131702572107315,
+ "learning_rate": 0.00019441100113920076,
+ "loss": 1.893791913986206,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51567944250871,
+ "grad_norm": 0.05804680287837982,
+ "learning_rate": 0.00019416291562234192,
+ "loss": 1.8838505744934082,
+ "step": 2269
+ },
+ {
+ "epoch": 31.529616724738677,
+ "grad_norm": 0.05933694168925285,
+ "learning_rate": 0.00019391491779355608,
+ "loss": 1.897843360900879,
+ "step": 2270
+ },
+ {
+ "epoch": 31.54355400696864,
+ "grad_norm": 0.058537356555461884,
+ "learning_rate": 0.00019366700785265065,
+ "loss": 1.8826205730438232,
+ "step": 2271
+ },
+ {
+ "epoch": 31.557491289198605,
+ "grad_norm": 0.060081496834754944,
+ "learning_rate": 0.0001934191859993622,
+ "loss": 1.889431357383728,
+ "step": 2272
+ },
+ {
+ "epoch": 31.571428571428573,
+ "grad_norm": 0.05904357507824898,
+ "learning_rate": 0.00019317145243335642,
+ "loss": 1.8965020179748535,
+ "step": 2273
+ },
+ {
+ "epoch": 31.585365853658537,
+ "grad_norm": 0.05924354866147041,
+ "learning_rate": 0.0001929238073542276,
+ "loss": 1.894204020500183,
+ "step": 2274
+ },
+ {
+ "epoch": 31.5993031358885,
+ "grad_norm": 0.059307269752025604,
+ "learning_rate": 0.00019267625096149903,
+ "loss": 1.890209436416626,
+ "step": 2275
+ },
+ {
+ "epoch": 31.613240418118465,
+ "grad_norm": 0.05969221144914627,
+ "learning_rate": 0.00019242878345462227,
+ "loss": 1.8865056037902832,
+ "step": 2276
+ },
+ {
+ "epoch": 31.627177700348433,
+ "grad_norm": 0.060738109052181244,
+ "learning_rate": 0.00019218140503297765,
+ "loss": 1.8825939893722534,
+ "step": 2277
+ },
+ {
+ "epoch": 31.641114982578397,
+ "grad_norm": 0.06001656875014305,
+ "learning_rate": 0.00019193411589587332,
+ "loss": 1.8988475799560547,
+ "step": 2278
+ },
+ {
+ "epoch": 31.65505226480836,
+ "grad_norm": 0.061478059738874435,
+ "learning_rate": 0.00019168691624254564,
+ "loss": 1.9071214199066162,
+ "step": 2279
+ },
+ {
+ "epoch": 31.66898954703833,
+ "grad_norm": 0.06089174002408981,
+ "learning_rate": 0.00019143980627215915,
+ "loss": 1.8998515605926514,
+ "step": 2280
+ },
+ {
+ "epoch": 31.682926829268293,
+ "grad_norm": 0.05940974876284599,
+ "learning_rate": 0.00019119278618380555,
+ "loss": 1.8907361030578613,
+ "step": 2281
+ },
+ {
+ "epoch": 31.696864111498257,
+ "grad_norm": 0.06015758961439133,
+ "learning_rate": 0.00019094585617650476,
+ "loss": 1.9010521173477173,
+ "step": 2282
+ },
+ {
+ "epoch": 31.71080139372822,
+ "grad_norm": 0.06062757968902588,
+ "learning_rate": 0.00019069901644920366,
+ "loss": 1.899493932723999,
+ "step": 2283
+ },
+ {
+ "epoch": 31.72473867595819,
+ "grad_norm": 0.060489553958177567,
+ "learning_rate": 0.00019045226720077667,
+ "loss": 1.9095027446746826,
+ "step": 2284
+ },
+ {
+ "epoch": 31.738675958188153,
+ "grad_norm": 0.05949694290757179,
+ "learning_rate": 0.0001902056086300251,
+ "loss": 1.900972604751587,
+ "step": 2285
+ },
+ {
+ "epoch": 31.752613240418118,
+ "grad_norm": 0.06269118189811707,
+ "learning_rate": 0.0001899590409356773,
+ "loss": 1.891621708869934,
+ "step": 2286
+ },
+ {
+ "epoch": 31.766550522648085,
+ "grad_norm": 0.06033318489789963,
+ "learning_rate": 0.00018971256431638854,
+ "loss": 1.8954113721847534,
+ "step": 2287
+ },
+ {
+ "epoch": 31.78048780487805,
+ "grad_norm": 0.0604838989675045,
+ "learning_rate": 0.00018946617897074041,
+ "loss": 1.9047741889953613,
+ "step": 2288
+ },
+ {
+ "epoch": 31.794425087108014,
+ "grad_norm": 0.060087885707616806,
+ "learning_rate": 0.0001892198850972413,
+ "loss": 1.8912427425384521,
+ "step": 2289
+ },
+ {
+ "epoch": 31.808362369337978,
+ "grad_norm": 0.059592097997665405,
+ "learning_rate": 0.00018897368289432553,
+ "loss": 1.8874431848526,
+ "step": 2290
+ },
+ {
+ "epoch": 31.822299651567945,
+ "grad_norm": 0.06219378113746643,
+ "learning_rate": 0.00018872757256035403,
+ "loss": 1.8843477964401245,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83623693379791,
+ "grad_norm": 0.06010756641626358,
+ "learning_rate": 0.00018848155429361323,
+ "loss": 1.892686367034912,
+ "step": 2292
+ },
+ {
+ "epoch": 31.850174216027874,
+ "grad_norm": 0.061404358595609665,
+ "learning_rate": 0.00018823562829231567,
+ "loss": 1.907541275024414,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86411149825784,
+ "grad_norm": 0.060503825545310974,
+ "learning_rate": 0.0001879897947545996,
+ "loss": 1.891597032546997,
+ "step": 2294
+ },
+ {
+ "epoch": 31.878048780487806,
+ "grad_norm": 0.06197115406394005,
+ "learning_rate": 0.00018774405387852848,
+ "loss": 1.8968595266342163,
+ "step": 2295
+ },
+ {
+ "epoch": 31.89198606271777,
+ "grad_norm": 0.06244310364127159,
+ "learning_rate": 0.00018749840586209144,
+ "loss": 1.903473138809204,
+ "step": 2296
+ },
+ {
+ "epoch": 31.905923344947734,
+ "grad_norm": 0.060648202896118164,
+ "learning_rate": 0.00018725285090320245,
+ "loss": 1.9044740200042725,
+ "step": 2297
+ },
+ {
+ "epoch": 31.9198606271777,
+ "grad_norm": 0.061299171298742294,
+ "learning_rate": 0.00018700738919970103,
+ "loss": 1.8890066146850586,
+ "step": 2298
+ },
+ {
+ "epoch": 31.933797909407666,
+ "grad_norm": 0.06204016134142876,
+ "learning_rate": 0.00018676202094935085,
+ "loss": 1.912150502204895,
+ "step": 2299
+ },
+ {
+ "epoch": 31.94773519163763,
+ "grad_norm": 0.06292745471000671,
+ "learning_rate": 0.000186516746349841,
+ "loss": 1.9133987426757812,
+ "step": 2300
+ },
+ {
+ "epoch": 31.961672473867594,
+ "grad_norm": 0.0633154958486557,
+ "learning_rate": 0.0001862715655987847,
+ "loss": 1.9106671810150146,
+ "step": 2301
+ },
+ {
+ "epoch": 31.975609756097562,
+ "grad_norm": 0.06269625574350357,
+ "learning_rate": 0.00018602647889371957,
+ "loss": 1.9090297222137451,
+ "step": 2302
+ },
+ {
+ "epoch": 31.989547038327526,
+ "grad_norm": 0.06060625612735748,
+ "learning_rate": 0.00018578148643210764,
+ "loss": 1.8874908685684204,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.05050232633948326,
+ "learning_rate": 0.00018553658841133483,
+ "loss": 1.4178190231323242,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 0.6386013627052307,
+ "eval_runtime": 40.9718,
+ "eval_samples_per_second": 59.602,
+ "eval_steps_per_second": 0.488,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01393728222997,
+ "grad_norm": 0.05701690912246704,
+ "learning_rate": 0.00018529178502871117,
+ "loss": 1.8772780895233154,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02787456445993,
+ "grad_norm": 0.05597634240984917,
+ "learning_rate": 0.0001850470764814702,
+ "loss": 1.8772567510604858,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041811846689896,
+ "grad_norm": 0.05631265789270401,
+ "learning_rate": 0.0001848024629667691,
+ "loss": 1.8660085201263428,
+ "step": 2307
+ },
+ {
+ "epoch": 32.055749128919864,
+ "grad_norm": 0.05604572966694832,
+ "learning_rate": 0.00018455794468168887,
+ "loss": 1.876168966293335,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069686411149824,
+ "grad_norm": 0.056123036891222,
+ "learning_rate": 0.00018431352182323322,
+ "loss": 1.8643467426300049,
+ "step": 2309
+ },
+ {
+ "epoch": 32.08362369337979,
+ "grad_norm": 0.05791968107223511,
+ "learning_rate": 0.0001840691945883293,
+ "loss": 1.866549015045166,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09756097560975,
+ "grad_norm": 0.05521981045603752,
+ "learning_rate": 0.00018382496317382715,
+ "loss": 1.8852781057357788,
+ "step": 2311
+ },
+ {
+ "epoch": 32.11149825783972,
+ "grad_norm": 0.055954866111278534,
+ "learning_rate": 0.0001835808277764997,
+ "loss": 1.8603641986846924,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12543554006969,
+ "grad_norm": 0.05931234359741211,
+ "learning_rate": 0.00018333678859304222,
+ "loss": 1.8619375228881836,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13937282229965,
+ "grad_norm": 0.05743015930056572,
+ "learning_rate": 0.00018309284582007274,
+ "loss": 1.8824023008346558,
+ "step": 2314
+ },
+ {
+ "epoch": 32.153310104529616,
+ "grad_norm": 0.056393202394247055,
+ "learning_rate": 0.0001828489996541318,
+ "loss": 1.8722079992294312,
+ "step": 2315
+ },
+ {
+ "epoch": 32.167247386759584,
+ "grad_norm": 0.05671187862753868,
+ "learning_rate": 0.00018260525029168147,
+ "loss": 1.8824248313903809,
+ "step": 2316
+ },
+ {
+ "epoch": 32.181184668989545,
+ "grad_norm": 0.05667170509696007,
+ "learning_rate": 0.00018236159792910648,
+ "loss": 1.882739782333374,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19512195121951,
+ "grad_norm": 0.05515167862176895,
+ "learning_rate": 0.00018211804276271304,
+ "loss": 1.8555771112442017,
+ "step": 2318
+ },
+ {
+ "epoch": 32.20905923344948,
+ "grad_norm": 0.05649147555232048,
+ "learning_rate": 0.00018187458498872914,
+ "loss": 1.8673419952392578,
+ "step": 2319
+ },
+ {
+ "epoch": 32.22299651567944,
+ "grad_norm": 0.058623913675546646,
+ "learning_rate": 0.00018163122480330433,
+ "loss": 1.8744664192199707,
+ "step": 2320
+ },
+ {
+ "epoch": 32.23693379790941,
+ "grad_norm": 0.055131133645772934,
+ "learning_rate": 0.00018138796240250955,
+ "loss": 1.8781135082244873,
+ "step": 2321
+ },
+ {
+ "epoch": 32.250871080139376,
+ "grad_norm": 0.06008831411600113,
+ "learning_rate": 0.00018114479798233686,
+ "loss": 1.8844444751739502,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26480836236934,
+ "grad_norm": 0.05755964294075966,
+ "learning_rate": 0.00018090173173869939,
+ "loss": 1.8790829181671143,
+ "step": 2323
+ },
+ {
+ "epoch": 32.278745644599304,
+ "grad_norm": 0.055803194642066956,
+ "learning_rate": 0.00018065876386743143,
+ "loss": 1.8642692565917969,
+ "step": 2324
+ },
+ {
+ "epoch": 32.292682926829265,
+ "grad_norm": 0.05710228905081749,
+ "learning_rate": 0.00018041589456428754,
+ "loss": 1.8659579753875732,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30662020905923,
+ "grad_norm": 0.05771065130829811,
+ "learning_rate": 0.0001801731240249434,
+ "loss": 1.8812355995178223,
+ "step": 2326
+ },
+ {
+ "epoch": 32.3205574912892,
+ "grad_norm": 0.05642757564783096,
+ "learning_rate": 0.00017993045244499463,
+ "loss": 1.8610783815383911,
+ "step": 2327
+ },
+ {
+ "epoch": 32.33449477351916,
+ "grad_norm": 0.05687689036130905,
+ "learning_rate": 0.00017968788001995742,
+ "loss": 1.875876545906067,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34843205574913,
+ "grad_norm": 0.057835984975099564,
+ "learning_rate": 0.000179445406945268,
+ "loss": 1.8674073219299316,
+ "step": 2329
+ },
+ {
+ "epoch": 32.362369337979096,
+ "grad_norm": 0.05800129100680351,
+ "learning_rate": 0.0001792030334162825,
+ "loss": 1.8708086013793945,
+ "step": 2330
+ },
+ {
+ "epoch": 32.37630662020906,
+ "grad_norm": 0.05744084715843201,
+ "learning_rate": 0.00017896075962827706,
+ "loss": 1.8869245052337646,
+ "step": 2331
+ },
+ {
+ "epoch": 32.390243902439025,
+ "grad_norm": 0.056703388690948486,
+ "learning_rate": 0.00017871858577644712,
+ "loss": 1.8688050508499146,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40418118466899,
+ "grad_norm": 0.06032602861523628,
+ "learning_rate": 0.0001784765120559078,
+ "loss": 1.873382568359375,
+ "step": 2333
+ },
+ {
+ "epoch": 32.41811846689895,
+ "grad_norm": 0.05810776352882385,
+ "learning_rate": 0.00017823453866169355,
+ "loss": 1.8624725341796875,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43205574912892,
+ "grad_norm": 0.060990724712610245,
+ "learning_rate": 0.00017799266578875808,
+ "loss": 1.8792011737823486,
+ "step": 2335
+ },
+ {
+ "epoch": 32.44599303135889,
+ "grad_norm": 0.05975554883480072,
+ "learning_rate": 0.00017775089363197371,
+ "loss": 1.8793275356292725,
+ "step": 2336
+ },
+ {
+ "epoch": 32.45993031358885,
+ "grad_norm": 0.05772533640265465,
+ "learning_rate": 0.00017750922238613198,
+ "loss": 1.889489769935608,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47386759581882,
+ "grad_norm": 0.05920211225748062,
+ "learning_rate": 0.00017726765224594342,
+ "loss": 1.8647074699401855,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48780487804878,
+ "grad_norm": 0.059745173901319504,
+ "learning_rate": 0.0001770261834060363,
+ "loss": 1.8649309873580933,
+ "step": 2339
+ },
+ {
+ "epoch": 32.501742160278745,
+ "grad_norm": 0.05933282524347305,
+ "learning_rate": 0.00017678481606095786,
+ "loss": 1.8936223983764648,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51567944250871,
+ "grad_norm": 0.062407393008470535,
+ "learning_rate": 0.0001765435504051734,
+ "loss": 1.8831725120544434,
+ "step": 2341
+ },
+ {
+ "epoch": 32.52961672473867,
+ "grad_norm": 0.061185337603092194,
+ "learning_rate": 0.0001763023866330664,
+ "loss": 1.8766106367111206,
+ "step": 2342
+ },
+ {
+ "epoch": 32.54355400696864,
+ "grad_norm": 0.06082421913743019,
+ "learning_rate": 0.00017606132493893793,
+ "loss": 1.880276083946228,
+ "step": 2343
+ },
+ {
+ "epoch": 32.55749128919861,
+ "grad_norm": 0.05900120362639427,
+ "learning_rate": 0.00017582036551700724,
+ "loss": 1.8696876764297485,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57142857142857,
+ "grad_norm": 0.06146271899342537,
+ "learning_rate": 0.00017557950856141085,
+ "loss": 1.864588737487793,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58536585365854,
+ "grad_norm": 0.05976773053407669,
+ "learning_rate": 0.00017533875426620286,
+ "loss": 1.8771488666534424,
+ "step": 2346
+ },
+ {
+ "epoch": 32.599303135888505,
+ "grad_norm": 0.05808904021978378,
+ "learning_rate": 0.00017509810282535497,
+ "loss": 1.890838623046875,
+ "step": 2347
+ },
+ {
+ "epoch": 32.613240418118465,
+ "grad_norm": 0.06149672716856003,
+ "learning_rate": 0.00017485755443275513,
+ "loss": 1.8787143230438232,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62717770034843,
+ "grad_norm": 0.059028271585702896,
+ "learning_rate": 0.0001746171092822094,
+ "loss": 1.8875718116760254,
+ "step": 2349
+ },
+ {
+ "epoch": 32.641114982578394,
+ "grad_norm": 0.062163304537534714,
+ "learning_rate": 0.00017437676756743974,
+ "loss": 1.86748206615448,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65505226480836,
+ "grad_norm": 0.06221195310354233,
+ "learning_rate": 0.00017413652948208524,
+ "loss": 1.8742306232452393,
+ "step": 2351
+ },
+ {
+ "epoch": 32.66898954703833,
+ "grad_norm": 0.060461003333330154,
+ "learning_rate": 0.00017389639521970136,
+ "loss": 1.885633945465088,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68292682926829,
+ "grad_norm": 0.06130106747150421,
+ "learning_rate": 0.00017365636497376004,
+ "loss": 1.8827314376831055,
+ "step": 2353
+ },
+ {
+ "epoch": 32.69686411149826,
+ "grad_norm": 0.06044832617044449,
+ "learning_rate": 0.00017341643893764927,
+ "loss": 1.8724548816680908,
+ "step": 2354
+ },
+ {
+ "epoch": 32.710801393728225,
+ "grad_norm": 0.06029713526368141,
+ "learning_rate": 0.000173176617304673,
+ "loss": 1.8745346069335938,
+ "step": 2355
+ },
+ {
+ "epoch": 32.724738675958186,
+ "grad_norm": 0.06287439167499542,
+ "learning_rate": 0.00017293690026805124,
+ "loss": 1.8879387378692627,
+ "step": 2356
+ },
+ {
+ "epoch": 32.73867595818815,
+ "grad_norm": 0.059636496007442474,
+ "learning_rate": 0.00017269728802091975,
+ "loss": 1.882830262184143,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75261324041812,
+ "grad_norm": 0.06590992957353592,
+ "learning_rate": 0.00017245778075632975,
+ "loss": 1.882242202758789,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76655052264808,
+ "grad_norm": 0.06283026933670044,
+ "learning_rate": 0.00017221837866724792,
+ "loss": 1.8864011764526367,
+ "step": 2359
+ },
+ {
+ "epoch": 32.78048780487805,
+ "grad_norm": 0.061717696487903595,
+ "learning_rate": 0.00017197908194655615,
+ "loss": 1.8815748691558838,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79442508710802,
+ "grad_norm": 0.06423293799161911,
+ "learning_rate": 0.00017173989078705175,
+ "loss": 1.8824026584625244,
+ "step": 2361
+ },
+ {
+ "epoch": 32.80836236933798,
+ "grad_norm": 0.06351540237665176,
+ "learning_rate": 0.0001715008053814463,
+ "loss": 1.9000264406204224,
+ "step": 2362
+ },
+ {
+ "epoch": 32.822299651567945,
+ "grad_norm": 0.060434527695178986,
+ "learning_rate": 0.00017126182592236682,
+ "loss": 1.8875732421875,
+ "step": 2363
+ },
+ {
+ "epoch": 32.836236933797906,
+ "grad_norm": 0.062305059283971786,
+ "learning_rate": 0.00017102295260235472,
+ "loss": 1.8695244789123535,
+ "step": 2364
+ },
+ {
+ "epoch": 32.850174216027874,
+ "grad_norm": 0.05981886014342308,
+ "learning_rate": 0.0001707841856138659,
+ "loss": 1.8885269165039062,
+ "step": 2365
+ },
+ {
+ "epoch": 32.86411149825784,
+ "grad_norm": 0.0642906129360199,
+ "learning_rate": 0.00017054552514927061,
+ "loss": 1.8818528652191162,
+ "step": 2366
+ },
+ {
+ "epoch": 32.8780487804878,
+ "grad_norm": 0.06316667050123215,
+ "learning_rate": 0.00017030697140085334,
+ "loss": 1.8994319438934326,
+ "step": 2367
+ },
+ {
+ "epoch": 32.89198606271777,
+ "grad_norm": 0.058260831981897354,
+ "learning_rate": 0.0001700685245608126,
+ "loss": 1.8908791542053223,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90592334494774,
+ "grad_norm": 0.062237389385700226,
+ "learning_rate": 0.00016983018482126047,
+ "loss": 1.8753025531768799,
+ "step": 2369
+ },
+ {
+ "epoch": 32.9198606271777,
+ "grad_norm": 0.05972493812441826,
+ "learning_rate": 0.00016959195237422305,
+ "loss": 1.899425745010376,
+ "step": 2370
+ },
+ {
+ "epoch": 32.933797909407666,
+ "grad_norm": 0.058846332132816315,
+ "learning_rate": 0.00016935382741164,
+ "loss": 1.8934335708618164,
+ "step": 2371
+ },
+ {
+ "epoch": 32.94773519163763,
+ "grad_norm": 0.06333424896001816,
+ "learning_rate": 0.00016911581012536424,
+ "loss": 1.8951395750045776,
+ "step": 2372
+ },
+ {
+ "epoch": 32.961672473867594,
+ "grad_norm": 0.06063693389296532,
+ "learning_rate": 0.0001688779007071622,
+ "loss": 1.881688117980957,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97560975609756,
+ "grad_norm": 0.060873813927173615,
+ "learning_rate": 0.00016864009934871275,
+ "loss": 1.8836770057678223,
+ "step": 2374
+ },
+ {
+ "epoch": 32.98954703832753,
+ "grad_norm": 0.060378748923540115,
+ "learning_rate": 0.0001684024062416087,
+ "loss": 1.886624813079834,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.04906065762042999,
+ "learning_rate": 0.0001681648215773546,
+ "loss": 1.4146361351013184,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 0.6401442885398865,
+ "eval_runtime": 41.0841,
+ "eval_samples_per_second": 59.439,
+ "eval_steps_per_second": 0.487,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01393728222997,
+ "grad_norm": 0.05557359382510185,
+ "learning_rate": 0.0001679273455473684,
+ "loss": 1.8605964183807373,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02787456445993,
+ "grad_norm": 0.055585991591215134,
+ "learning_rate": 0.0001676899783429802,
+ "loss": 1.8456363677978516,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041811846689896,
+ "grad_norm": 0.05610295757651329,
+ "learning_rate": 0.00016745272015543242,
+ "loss": 1.847888469696045,
+ "step": 2379
+ },
+ {
+ "epoch": 33.055749128919864,
+ "grad_norm": 0.05532662197947502,
+ "learning_rate": 0.0001672155711758799,
+ "loss": 1.850597620010376,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069686411149824,
+ "grad_norm": 0.05585319921374321,
+ "learning_rate": 0.00016697853159538896,
+ "loss": 1.8694689273834229,
+ "step": 2381
+ },
+ {
+ "epoch": 33.08362369337979,
+ "grad_norm": 0.05658521130681038,
+ "learning_rate": 0.0001667416016049383,
+ "loss": 1.8533908128738403,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09756097560975,
+ "grad_norm": 0.0558413490653038,
+ "learning_rate": 0.0001665047813954181,
+ "loss": 1.8735582828521729,
+ "step": 2383
+ },
+ {
+ "epoch": 33.11149825783972,
+ "grad_norm": 0.05637028068304062,
+ "learning_rate": 0.00016626807115763023,
+ "loss": 1.8585509061813354,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12543554006969,
+ "grad_norm": 0.055854618549346924,
+ "learning_rate": 0.00016603147108228754,
+ "loss": 1.8539409637451172,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13937282229965,
+ "grad_norm": 0.055990077555179596,
+ "learning_rate": 0.00016579498136001485,
+ "loss": 1.85886812210083,
+ "step": 2386
+ },
+ {
+ "epoch": 33.153310104529616,
+ "grad_norm": 0.05781752988696098,
+ "learning_rate": 0.00016555860218134738,
+ "loss": 1.8603525161743164,
+ "step": 2387
+ },
+ {
+ "epoch": 33.167247386759584,
+ "grad_norm": 0.055855557322502136,
+ "learning_rate": 0.00016532233373673164,
+ "loss": 1.861933946609497,
+ "step": 2388
+ },
+ {
+ "epoch": 33.181184668989545,
+ "grad_norm": 0.05729927122592926,
+ "learning_rate": 0.00016508617621652488,
+ "loss": 1.86094069480896,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19512195121951,
+ "grad_norm": 0.056495532393455505,
+ "learning_rate": 0.00016485012981099488,
+ "loss": 1.848013997077942,
+ "step": 2390
+ },
+ {
+ "epoch": 33.20905923344948,
+ "grad_norm": 0.0571073479950428,
+ "learning_rate": 0.0001646141947103201,
+ "loss": 1.8558173179626465,
+ "step": 2391
+ },
+ {
+ "epoch": 33.22299651567944,
+ "grad_norm": 0.056066226214170456,
+ "learning_rate": 0.0001643783711045889,
+ "loss": 1.8723382949829102,
+ "step": 2392
+ },
+ {
+ "epoch": 33.23693379790941,
+ "grad_norm": 0.05570382624864578,
+ "learning_rate": 0.0001641426591838005,
+ "loss": 1.864349365234375,
+ "step": 2393
+ },
+ {
+ "epoch": 33.250871080139376,
+ "grad_norm": 0.054912492632865906,
+ "learning_rate": 0.00016390705913786344,
+ "loss": 1.8678479194641113,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26480836236934,
+ "grad_norm": 0.05622183158993721,
+ "learning_rate": 0.00016367157115659646,
+ "loss": 1.8648667335510254,
+ "step": 2395
+ },
+ {
+ "epoch": 33.278745644599304,
+ "grad_norm": 0.05719740688800812,
+ "learning_rate": 0.00016343619542972808,
+ "loss": 1.850623607635498,
+ "step": 2396
+ },
+ {
+ "epoch": 33.292682926829265,
+ "grad_norm": 0.056594349443912506,
+ "learning_rate": 0.0001632009321468962,
+ "loss": 1.8707754611968994,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30662020905923,
+ "grad_norm": 0.05596613511443138,
+ "learning_rate": 0.00016296578149764833,
+ "loss": 1.8580527305603027,
+ "step": 2398
+ },
+ {
+ "epoch": 33.3205574912892,
+ "grad_norm": 0.05675627663731575,
+ "learning_rate": 0.00016273074367144087,
+ "loss": 1.8687434196472168,
+ "step": 2399
+ },
+ {
+ "epoch": 33.33449477351916,
+ "grad_norm": 0.05572087690234184,
+ "learning_rate": 0.00016249581885763968,
+ "loss": 1.8603391647338867,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34843205574913,
+ "grad_norm": 0.05837463587522507,
+ "learning_rate": 0.0001622610072455194,
+ "loss": 1.8638341426849365,
+ "step": 2401
+ },
+ {
+ "epoch": 33.362369337979096,
+ "grad_norm": 0.056136030703783035,
+ "learning_rate": 0.00016202630902426355,
+ "loss": 1.853350281715393,
+ "step": 2402
+ },
+ {
+ "epoch": 33.37630662020906,
+ "grad_norm": 0.0555158406496048,
+ "learning_rate": 0.00016179172438296423,
+ "loss": 1.8775548934936523,
+ "step": 2403
+ },
+ {
+ "epoch": 33.390243902439025,
+ "grad_norm": 0.0595950111746788,
+ "learning_rate": 0.00016155725351062205,
+ "loss": 1.8578253984451294,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40418118466899,
+ "grad_norm": 0.05565760284662247,
+ "learning_rate": 0.00016132289659614605,
+ "loss": 1.8508214950561523,
+ "step": 2405
+ },
+ {
+ "epoch": 33.41811846689895,
+ "grad_norm": 0.058832235634326935,
+ "learning_rate": 0.00016108865382835318,
+ "loss": 1.8614635467529297,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43205574912892,
+ "grad_norm": 0.0585586316883564,
+ "learning_rate": 0.00016085452539596864,
+ "loss": 1.8617370128631592,
+ "step": 2407
+ },
+ {
+ "epoch": 33.44599303135889,
+ "grad_norm": 0.05951899662613869,
+ "learning_rate": 0.0001606205114876256,
+ "loss": 1.8648651838302612,
+ "step": 2408
+ },
+ {
+ "epoch": 33.45993031358885,
+ "grad_norm": 0.057204652577638626,
+ "learning_rate": 0.00016038661229186472,
+ "loss": 1.8577271699905396,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47386759581882,
+ "grad_norm": 0.05844349041581154,
+ "learning_rate": 0.00016015282799713453,
+ "loss": 1.8560724258422852,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48780487804878,
+ "grad_norm": 0.05761651322245598,
+ "learning_rate": 0.0001599191587917905,
+ "loss": 1.863309383392334,
+ "step": 2411
+ },
+ {
+ "epoch": 33.501742160278745,
+ "grad_norm": 0.05818328261375427,
+ "learning_rate": 0.0001596856048640961,
+ "loss": 1.8798553943634033,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51567944250871,
+ "grad_norm": 0.05746658146381378,
+ "learning_rate": 0.00015945216640222116,
+ "loss": 1.8628616333007812,
+ "step": 2413
+ },
+ {
+ "epoch": 33.52961672473867,
+ "grad_norm": 0.057656314224004745,
+ "learning_rate": 0.000159218843594243,
+ "loss": 1.8699405193328857,
+ "step": 2414
+ },
+ {
+ "epoch": 33.54355400696864,
+ "grad_norm": 0.0586664155125618,
+ "learning_rate": 0.0001589856366281455,
+ "loss": 1.8481391668319702,
+ "step": 2415
+ },
+ {
+ "epoch": 33.55749128919861,
+ "grad_norm": 0.05769997090101242,
+ "learning_rate": 0.0001587525456918194,
+ "loss": 1.8612866401672363,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57142857142857,
+ "grad_norm": 0.0598456896841526,
+ "learning_rate": 0.0001585195709730619,
+ "loss": 1.8580410480499268,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58536585365854,
+ "grad_norm": 0.05761442333459854,
+ "learning_rate": 0.00015828671265957623,
+ "loss": 1.874891996383667,
+ "step": 2418
+ },
+ {
+ "epoch": 33.599303135888505,
+ "grad_norm": 0.06176742538809776,
+ "learning_rate": 0.00015805397093897255,
+ "loss": 1.8660908937454224,
+ "step": 2419
+ },
+ {
+ "epoch": 33.613240418118465,
+ "grad_norm": 0.0588943213224411,
+ "learning_rate": 0.00015782134599876633,
+ "loss": 1.8718886375427246,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62717770034843,
+ "grad_norm": 0.06092666834592819,
+ "learning_rate": 0.00015758883802637953,
+ "loss": 1.8634107112884521,
+ "step": 2421
+ },
+ {
+ "epoch": 33.641114982578394,
+ "grad_norm": 0.05774684250354767,
+ "learning_rate": 0.00015735644720913923,
+ "loss": 1.868361234664917,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65505226480836,
+ "grad_norm": 0.05932781100273132,
+ "learning_rate": 0.00015712417373427897,
+ "loss": 1.867889642715454,
+ "step": 2423
+ },
+ {
+ "epoch": 33.66898954703833,
+ "grad_norm": 0.058603718876838684,
+ "learning_rate": 0.00015689201778893713,
+ "loss": 1.8793346881866455,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68292682926829,
+ "grad_norm": 0.05870316922664642,
+ "learning_rate": 0.00015665997956015743,
+ "loss": 1.8717231750488281,
+ "step": 2425
+ },
+ {
+ "epoch": 33.69686411149826,
+ "grad_norm": 0.05805273354053497,
+ "learning_rate": 0.00015642805923488895,
+ "loss": 1.8479297161102295,
+ "step": 2426
+ },
+ {
+ "epoch": 33.710801393728225,
+ "grad_norm": 0.059018444269895554,
+ "learning_rate": 0.00015619625699998585,
+ "loss": 1.8711276054382324,
+ "step": 2427
+ },
+ {
+ "epoch": 33.724738675958186,
+ "grad_norm": 0.05927985534071922,
+ "learning_rate": 0.00015596457304220692,
+ "loss": 1.8653428554534912,
+ "step": 2428
+ },
+ {
+ "epoch": 33.73867595818815,
+ "grad_norm": 0.06085766851902008,
+ "learning_rate": 0.00015573300754821547,
+ "loss": 1.8724851608276367,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75261324041812,
+ "grad_norm": 0.05974132940173149,
+ "learning_rate": 0.00015550156070458002,
+ "loss": 1.8459011316299438,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76655052264808,
+ "grad_norm": 0.05851755291223526,
+ "learning_rate": 0.0001552702326977731,
+ "loss": 1.863109827041626,
+ "step": 2431
+ },
+ {
+ "epoch": 33.78048780487805,
+ "grad_norm": 0.06079421192407608,
+ "learning_rate": 0.00015503902371417122,
+ "loss": 1.8687033653259277,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79442508710802,
+ "grad_norm": 0.057930078357458115,
+ "learning_rate": 0.00015480793394005547,
+ "loss": 1.8653396368026733,
+ "step": 2433
+ },
+ {
+ "epoch": 33.80836236933798,
+ "grad_norm": 0.059440065175294876,
+ "learning_rate": 0.0001545769635616107,
+ "loss": 1.8569486141204834,
+ "step": 2434
+ },
+ {
+ "epoch": 33.822299651567945,
+ "grad_norm": 0.05835104361176491,
+ "learning_rate": 0.0001543461127649256,
+ "loss": 1.8828535079956055,
+ "step": 2435
+ },
+ {
+ "epoch": 33.836236933797906,
+ "grad_norm": 0.06108623370528221,
+ "learning_rate": 0.00015411538173599231,
+ "loss": 1.8749947547912598,
+ "step": 2436
+ },
+ {
+ "epoch": 33.850174216027874,
+ "grad_norm": 0.059921979904174805,
+ "learning_rate": 0.00015388477066070674,
+ "loss": 1.8684660196304321,
+ "step": 2437
+ },
+ {
+ "epoch": 33.86411149825784,
+ "grad_norm": 0.059964295476675034,
+ "learning_rate": 0.00015365427972486827,
+ "loss": 1.862084150314331,
+ "step": 2438
+ },
+ {
+ "epoch": 33.8780487804878,
+ "grad_norm": 0.05888533219695091,
+ "learning_rate": 0.00015342390911417904,
+ "loss": 1.85898756980896,
+ "step": 2439
+ },
+ {
+ "epoch": 33.89198606271777,
+ "grad_norm": 0.058946020901203156,
+ "learning_rate": 0.00015319365901424452,
+ "loss": 1.872513771057129,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90592334494774,
+ "grad_norm": 0.05947734788060188,
+ "learning_rate": 0.0001529635296105732,
+ "loss": 1.8610495328903198,
+ "step": 2441
+ },
+ {
+ "epoch": 33.9198606271777,
+ "grad_norm": 0.05859588086605072,
+ "learning_rate": 0.0001527335210885761,
+ "loss": 1.8842005729675293,
+ "step": 2442
+ },
+ {
+ "epoch": 33.933797909407666,
+ "grad_norm": 0.05818016827106476,
+ "learning_rate": 0.00015250363363356685,
+ "loss": 1.8715083599090576,
+ "step": 2443
+ },
+ {
+ "epoch": 33.94773519163763,
+ "grad_norm": 0.05960588902235031,
+ "learning_rate": 0.00015227386743076155,
+ "loss": 1.881840467453003,
+ "step": 2444
+ },
+ {
+ "epoch": 33.961672473867594,
+ "grad_norm": 0.059536516666412354,
+ "learning_rate": 0.00015204422266527911,
+ "loss": 1.8734736442565918,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97560975609756,
+ "grad_norm": 0.05922378972172737,
+ "learning_rate": 0.00015181469952213968,
+ "loss": 1.8569964170455933,
+ "step": 2446
+ },
+ {
+ "epoch": 33.98954703832753,
+ "grad_norm": 0.058573752641677856,
+ "learning_rate": 0.0001515852981862663,
+ "loss": 1.8710730075836182,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.047325484454631805,
+ "learning_rate": 0.00015135601884248303,
+ "loss": 1.4022152423858643,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 0.6414204835891724,
+ "eval_runtime": 39.9489,
+ "eval_samples_per_second": 61.128,
+ "eval_steps_per_second": 0.501,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01393728222997,
+ "grad_norm": 0.05358244478702545,
+ "learning_rate": 0.0001511268616755166,
+ "loss": 1.8384610414505005,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02787456445993,
+ "grad_norm": 0.0558239221572876,
+ "learning_rate": 0.00015089782686999445,
+ "loss": 1.8538508415222168,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041811846689896,
+ "grad_norm": 0.05390622094273567,
+ "learning_rate": 0.00015066891461044588,
+ "loss": 1.8460195064544678,
+ "step": 2451
+ },
+ {
+ "epoch": 34.055749128919864,
+ "grad_norm": 0.056224383413791656,
+ "learning_rate": 0.00015044012508130138,
+ "loss": 1.8416059017181396,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069686411149824,
+ "grad_norm": 0.053925771266222,
+ "learning_rate": 0.00015021145846689254,
+ "loss": 1.8532803058624268,
+ "step": 2453
+ },
+ {
+ "epoch": 34.08362369337979,
+ "grad_norm": 0.05750057101249695,
+ "learning_rate": 0.00014998291495145206,
+ "loss": 1.847919225692749,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09756097560975,
+ "grad_norm": 0.05499042198061943,
+ "learning_rate": 0.00014975449471911294,
+ "loss": 1.8450297117233276,
+ "step": 2455
+ },
+ {
+ "epoch": 34.11149825783972,
+ "grad_norm": 0.0544770173728466,
+ "learning_rate": 0.0001495261979539097,
+ "loss": 1.8482413291931152,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12543554006969,
+ "grad_norm": 0.05656695365905762,
+ "learning_rate": 0.0001492980248397766,
+ "loss": 1.860766053199768,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13937282229965,
+ "grad_norm": 0.054617058485746384,
+ "learning_rate": 0.00014906997556054865,
+ "loss": 1.8398511409759521,
+ "step": 2458
+ },
+ {
+ "epoch": 34.153310104529616,
+ "grad_norm": 0.05470099300146103,
+ "learning_rate": 0.00014884205029996113,
+ "loss": 1.8501665592193604,
+ "step": 2459
+ },
+ {
+ "epoch": 34.167247386759584,
+ "grad_norm": 0.05437156930565834,
+ "learning_rate": 0.00014861424924164925,
+ "loss": 1.8514611721038818,
+ "step": 2460
+ },
+ {
+ "epoch": 34.181184668989545,
+ "grad_norm": 0.05481892451643944,
+ "learning_rate": 0.00014838657256914832,
+ "loss": 1.8424533605575562,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19512195121951,
+ "grad_norm": 0.055016402155160904,
+ "learning_rate": 0.0001481590204658931,
+ "loss": 1.823396921157837,
+ "step": 2462
+ },
+ {
+ "epoch": 34.20905923344948,
+ "grad_norm": 0.05608086287975311,
+ "learning_rate": 0.00014793159311521833,
+ "loss": 1.8525217771530151,
+ "step": 2463
+ },
+ {
+ "epoch": 34.22299651567944,
+ "grad_norm": 0.055000413209199905,
+ "learning_rate": 0.00014770429070035812,
+ "loss": 1.8570690155029297,
+ "step": 2464
+ },
+ {
+ "epoch": 34.23693379790941,
+ "grad_norm": 0.05437581241130829,
+ "learning_rate": 0.00014747711340444601,
+ "loss": 1.8471980094909668,
+ "step": 2465
+ },
+ {
+ "epoch": 34.250871080139376,
+ "grad_norm": 0.0553814135491848,
+ "learning_rate": 0.00014725006141051434,
+ "loss": 1.8384971618652344,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26480836236934,
+ "grad_norm": 0.053538259118795395,
+ "learning_rate": 0.00014702313490149512,
+ "loss": 1.8536994457244873,
+ "step": 2467
+ },
+ {
+ "epoch": 34.278745644599304,
+ "grad_norm": 0.05509314313530922,
+ "learning_rate": 0.00014679633406021892,
+ "loss": 1.8386962413787842,
+ "step": 2468
+ },
+ {
+ "epoch": 34.292682926829265,
+ "grad_norm": 0.054552506655454636,
+ "learning_rate": 0.00014656965906941494,
+ "loss": 1.8502833843231201,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30662020905923,
+ "grad_norm": 0.054326511919498444,
+ "learning_rate": 0.00014634311011171112,
+ "loss": 1.8350508213043213,
+ "step": 2470
+ },
+ {
+ "epoch": 34.3205574912892,
+ "grad_norm": 0.05458610877394676,
+ "learning_rate": 0.00014611668736963396,
+ "loss": 1.8488143682479858,
+ "step": 2471
+ },
+ {
+ "epoch": 34.33449477351916,
+ "grad_norm": 0.05537666752934456,
+ "learning_rate": 0.00014589039102560822,
+ "loss": 1.8561567068099976,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34843205574913,
+ "grad_norm": 0.056333623826503754,
+ "learning_rate": 0.00014566422126195656,
+ "loss": 1.8610734939575195,
+ "step": 2473
+ },
+ {
+ "epoch": 34.362369337979096,
+ "grad_norm": 0.053708501160144806,
+ "learning_rate": 0.00014543817826089992,
+ "loss": 1.8342688083648682,
+ "step": 2474
+ },
+ {
+ "epoch": 34.37630662020906,
+ "grad_norm": 0.05496754124760628,
+ "learning_rate": 0.0001452122622045574,
+ "loss": 1.8592373132705688,
+ "step": 2475
+ },
+ {
+ "epoch": 34.390243902439025,
+ "grad_norm": 0.05445944517850876,
+ "learning_rate": 0.00014498647327494506,
+ "loss": 1.8551034927368164,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40418118466899,
+ "grad_norm": 0.055870212614536285,
+ "learning_rate": 0.00014476081165397715,
+ "loss": 1.8510769605636597,
+ "step": 2477
+ },
+ {
+ "epoch": 34.41811846689895,
+ "grad_norm": 0.05536561831831932,
+ "learning_rate": 0.00014453527752346517,
+ "loss": 1.8463027477264404,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43205574912892,
+ "grad_norm": 0.054308436810970306,
+ "learning_rate": 0.00014430987106511795,
+ "loss": 1.8470419645309448,
+ "step": 2479
+ },
+ {
+ "epoch": 34.44599303135889,
+ "grad_norm": 0.056683722883462906,
+ "learning_rate": 0.00014408459246054126,
+ "loss": 1.8605597019195557,
+ "step": 2480
+ },
+ {
+ "epoch": 34.45993031358885,
+ "grad_norm": 0.05601316690444946,
+ "learning_rate": 0.00014385944189123794,
+ "loss": 1.847717046737671,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47386759581882,
+ "grad_norm": 0.05449012294411659,
+ "learning_rate": 0.00014363441953860805,
+ "loss": 1.8464031219482422,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48780487804878,
+ "grad_norm": 0.054641272872686386,
+ "learning_rate": 0.00014340952558394778,
+ "loss": 1.832083821296692,
+ "step": 2483
+ },
+ {
+ "epoch": 34.501742160278745,
+ "grad_norm": 0.05610528588294983,
+ "learning_rate": 0.00014318476020845018,
+ "loss": 1.8700275421142578,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51567944250871,
+ "grad_norm": 0.0551300011575222,
+ "learning_rate": 0.0001429601235932047,
+ "loss": 1.8514257669448853,
+ "step": 2485
+ },
+ {
+ "epoch": 34.52961672473867,
+ "grad_norm": 0.056593794375658035,
+ "learning_rate": 0.00014273561591919705,
+ "loss": 1.8480379581451416,
+ "step": 2486
+ },
+ {
+ "epoch": 34.54355400696864,
+ "grad_norm": 0.055673424154520035,
+ "learning_rate": 0.00014251123736730884,
+ "loss": 1.838927984237671,
+ "step": 2487
+ },
+ {
+ "epoch": 34.55749128919861,
+ "grad_norm": 0.05493558943271637,
+ "learning_rate": 0.00014228698811831787,
+ "loss": 1.8383840322494507,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57142857142857,
+ "grad_norm": 0.05680295452475548,
+ "learning_rate": 0.0001420628683528978,
+ "loss": 1.8514118194580078,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58536585365854,
+ "grad_norm": 0.057038743048906326,
+ "learning_rate": 0.00014183887825161779,
+ "loss": 1.8620257377624512,
+ "step": 2490
+ },
+ {
+ "epoch": 34.599303135888505,
+ "grad_norm": 0.05735227093100548,
+ "learning_rate": 0.0001416150179949428,
+ "loss": 1.8413968086242676,
+ "step": 2491
+ },
+ {
+ "epoch": 34.613240418118465,
+ "grad_norm": 0.058245740830898285,
+ "learning_rate": 0.0001413912877632326,
+ "loss": 1.8599750995635986,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62717770034843,
+ "grad_norm": 0.056244585663080215,
+ "learning_rate": 0.00014116768773674306,
+ "loss": 1.8527486324310303,
+ "step": 2493
+ },
+ {
+ "epoch": 34.641114982578394,
+ "grad_norm": 0.05616258084774017,
+ "learning_rate": 0.00014094421809562438,
+ "loss": 1.84233558177948,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65505226480836,
+ "grad_norm": 0.057303108274936676,
+ "learning_rate": 0.00014072087901992209,
+ "loss": 1.865498661994934,
+ "step": 2495
+ },
+ {
+ "epoch": 34.66898954703833,
+ "grad_norm": 0.05895329639315605,
+ "learning_rate": 0.00014049767068957645,
+ "loss": 1.8521645069122314,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68292682926829,
+ "grad_norm": 0.05679212138056755,
+ "learning_rate": 0.00014027459328442242,
+ "loss": 1.8539199829101562,
+ "step": 2497
+ },
+ {
+ "epoch": 34.69686411149826,
+ "grad_norm": 0.05897561088204384,
+ "learning_rate": 0.00014005164698418948,
+ "loss": 1.8534815311431885,
+ "step": 2498
+ },
+ {
+ "epoch": 34.710801393728225,
+ "grad_norm": 0.05891786143183708,
+ "learning_rate": 0.0001398288319685013,
+ "loss": 1.845531702041626,
+ "step": 2499
+ },
+ {
+ "epoch": 34.724738675958186,
+ "grad_norm": 0.059177134186029434,
+ "learning_rate": 0.00013960614841687595,
+ "loss": 1.8502740859985352,
+ "step": 2500
+ },
+ {
+ "epoch": 34.73867595818815,
+ "grad_norm": 0.05862346291542053,
+ "learning_rate": 0.00013938359650872558,
+ "loss": 1.8657329082489014,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75261324041812,
+ "grad_norm": 0.061109986156225204,
+ "learning_rate": 0.00013916117642335623,
+ "loss": 1.8452383279800415,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76655052264808,
+ "grad_norm": 0.058431629091501236,
+ "learning_rate": 0.00013893888833996774,
+ "loss": 1.8575074672698975,
+ "step": 2503
+ },
+ {
+ "epoch": 34.78048780487805,
+ "grad_norm": 0.05970519408583641,
+ "learning_rate": 0.00013871673243765362,
+ "loss": 1.8476191759109497,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79442508710802,
+ "grad_norm": 0.05823173746466637,
+ "learning_rate": 0.00013849470889540098,
+ "loss": 1.8554785251617432,
+ "step": 2505
+ },
+ {
+ "epoch": 34.80836236933798,
+ "grad_norm": 0.060084082186222076,
+ "learning_rate": 0.00013827281789208996,
+ "loss": 1.8471784591674805,
+ "step": 2506
+ },
+ {
+ "epoch": 34.822299651567945,
+ "grad_norm": 0.0589512400329113,
+ "learning_rate": 0.00013805105960649423,
+ "loss": 1.86192786693573,
+ "step": 2507
+ },
+ {
+ "epoch": 34.836236933797906,
+ "grad_norm": 0.06184513866901398,
+ "learning_rate": 0.00013782943421728042,
+ "loss": 1.8572009801864624,
+ "step": 2508
+ },
+ {
+ "epoch": 34.850174216027874,
+ "grad_norm": 0.05833953991532326,
+ "learning_rate": 0.00013760794190300827,
+ "loss": 1.8561553955078125,
+ "step": 2509
+ },
+ {
+ "epoch": 34.86411149825784,
+ "grad_norm": 0.061172984540462494,
+ "learning_rate": 0.00013738658284212973,
+ "loss": 1.8583234548568726,
+ "step": 2510
+ },
+ {
+ "epoch": 34.8780487804878,
+ "grad_norm": 0.0573224313557148,
+ "learning_rate": 0.00013716535721299016,
+ "loss": 1.8403494358062744,
+ "step": 2511
+ },
+ {
+ "epoch": 34.89198606271777,
+ "grad_norm": 0.06135294586420059,
+ "learning_rate": 0.0001369442651938271,
+ "loss": 1.869114875793457,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90592334494774,
+ "grad_norm": 0.05907885357737541,
+ "learning_rate": 0.00013672330696277014,
+ "loss": 1.8581886291503906,
+ "step": 2513
+ },
+ {
+ "epoch": 34.9198606271777,
+ "grad_norm": 0.06035003438591957,
+ "learning_rate": 0.00013650248269784143,
+ "loss": 1.8454012870788574,
+ "step": 2514
+ },
+ {
+ "epoch": 34.933797909407666,
+ "grad_norm": 0.060585640370845795,
+ "learning_rate": 0.00013628179257695508,
+ "loss": 1.8453309535980225,
+ "step": 2515
+ },
+ {
+ "epoch": 34.94773519163763,
+ "grad_norm": 0.06050759181380272,
+ "learning_rate": 0.00013606123677791727,
+ "loss": 1.8467144966125488,
+ "step": 2516
+ },
+ {
+ "epoch": 34.961672473867594,
+ "grad_norm": 0.06118239462375641,
+ "learning_rate": 0.00013584081547842565,
+ "loss": 1.8623638153076172,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97560975609756,
+ "grad_norm": 0.059003446251153946,
+ "learning_rate": 0.00013562052885606957,
+ "loss": 1.8540363311767578,
+ "step": 2518
+ },
+ {
+ "epoch": 34.98954703832753,
+ "grad_norm": 0.06078164279460907,
+ "learning_rate": 0.00013540037708833043,
+ "loss": 1.8434364795684814,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.04768386855721474,
+ "learning_rate": 0.00013518036035258012,
+ "loss": 1.3819507360458374,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 0.642964243888855,
+ "eval_runtime": 40.0846,
+ "eval_samples_per_second": 60.921,
+ "eval_steps_per_second": 0.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01393728222997,
+ "grad_norm": 0.05531901493668556,
+ "learning_rate": 0.0001349604788260823,
+ "loss": 1.8454805612564087,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02787456445993,
+ "grad_norm": 0.05443203076720238,
+ "learning_rate": 0.00013474073268599152,
+ "loss": 1.8266499042510986,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041811846689896,
+ "grad_norm": 0.05491992458701134,
+ "learning_rate": 0.00013452112210935339,
+ "loss": 1.839545488357544,
+ "step": 2523
+ },
+ {
+ "epoch": 35.055749128919864,
+ "grad_norm": 0.05408782139420509,
+ "learning_rate": 0.00013430164727310399,
+ "loss": 1.836404800415039,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069686411149824,
+ "grad_norm": 0.0541234128177166,
+ "learning_rate": 0.00013408230835407027,
+ "loss": 1.8264034986495972,
+ "step": 2525
+ },
+ {
+ "epoch": 35.08362369337979,
+ "grad_norm": 0.05504459887742996,
+ "learning_rate": 0.00013386310552896977,
+ "loss": 1.8485287427902222,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09756097560975,
+ "grad_norm": 0.05482470244169235,
+ "learning_rate": 0.00013364403897441016,
+ "loss": 1.8336119651794434,
+ "step": 2527
+ },
+ {
+ "epoch": 35.11149825783972,
+ "grad_norm": 0.05490747466683388,
+ "learning_rate": 0.00013342510886688944,
+ "loss": 1.8223836421966553,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12543554006969,
+ "grad_norm": 0.05421694368124008,
+ "learning_rate": 0.00013320631538279548,
+ "loss": 1.8309847116470337,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13937282229965,
+ "grad_norm": 0.05432746186852455,
+ "learning_rate": 0.00013298765869840658,
+ "loss": 1.8214728832244873,
+ "step": 2530
+ },
+ {
+ "epoch": 35.153310104529616,
+ "grad_norm": 0.05329402908682823,
+ "learning_rate": 0.00013276913898989013,
+ "loss": 1.8356881141662598,
+ "step": 2531
+ },
+ {
+ "epoch": 35.167247386759584,
+ "grad_norm": 0.05515717342495918,
+ "learning_rate": 0.00013255075643330369,
+ "loss": 1.8358293771743774,
+ "step": 2532
+ },
+ {
+ "epoch": 35.181184668989545,
+ "grad_norm": 0.05425635725259781,
+ "learning_rate": 0.00013233251120459407,
+ "loss": 1.8230032920837402,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19512195121951,
+ "grad_norm": 0.054918598383665085,
+ "learning_rate": 0.0001321144034795976,
+ "loss": 1.8427648544311523,
+ "step": 2534
+ },
+ {
+ "epoch": 35.20905923344948,
+ "grad_norm": 0.05530979484319687,
+ "learning_rate": 0.00013189643343403977,
+ "loss": 1.811009407043457,
+ "step": 2535
+ },
+ {
+ "epoch": 35.22299651567944,
+ "grad_norm": 0.05482922121882439,
+ "learning_rate": 0.00013167860124353476,
+ "loss": 1.8356122970581055,
+ "step": 2536
+ },
+ {
+ "epoch": 35.23693379790941,
+ "grad_norm": 0.05470896139740944,
+ "learning_rate": 0.00013146090708358657,
+ "loss": 1.8432835340499878,
+ "step": 2537
+ },
+ {
+ "epoch": 35.250871080139376,
+ "grad_norm": 0.05382648482918739,
+ "learning_rate": 0.00013124335112958704,
+ "loss": 1.8248438835144043,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26480836236934,
+ "grad_norm": 0.05452217534184456,
+ "learning_rate": 0.0001310259335568172,
+ "loss": 1.8385814428329468,
+ "step": 2539
+ },
+ {
+ "epoch": 35.278745644599304,
+ "grad_norm": 0.05492600053548813,
+ "learning_rate": 0.00013080865454044647,
+ "loss": 1.8321887254714966,
+ "step": 2540
+ },
+ {
+ "epoch": 35.292682926829265,
+ "grad_norm": 0.054316446185112,
+ "learning_rate": 0.00013059151425553263,
+ "loss": 1.8411835432052612,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30662020905923,
+ "grad_norm": 0.05490174889564514,
+ "learning_rate": 0.00013037451287702183,
+ "loss": 1.8350071907043457,
+ "step": 2542
+ },
+ {
+ "epoch": 35.3205574912892,
+ "grad_norm": 0.055120889097452164,
+ "learning_rate": 0.0001301576505797479,
+ "loss": 1.8385683298110962,
+ "step": 2543
+ },
+ {
+ "epoch": 35.33449477351916,
+ "grad_norm": 0.054340824484825134,
+ "learning_rate": 0.00012994092753843293,
+ "loss": 1.856696605682373,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34843205574913,
+ "grad_norm": 0.056243546307086945,
+ "learning_rate": 0.00012972434392768687,
+ "loss": 1.8386306762695312,
+ "step": 2545
+ },
+ {
+ "epoch": 35.362369337979096,
+ "grad_norm": 0.05486825481057167,
+ "learning_rate": 0.00012950789992200714,
+ "loss": 1.8275195360183716,
+ "step": 2546
+ },
+ {
+ "epoch": 35.37630662020906,
+ "grad_norm": 0.05750235170125961,
+ "learning_rate": 0.00012929159569577886,
+ "loss": 1.8266764879226685,
+ "step": 2547
+ },
+ {
+ "epoch": 35.390243902439025,
+ "grad_norm": 0.05539723485708237,
+ "learning_rate": 0.0001290754314232743,
+ "loss": 1.8288836479187012,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40418118466899,
+ "grad_norm": 0.056604526937007904,
+ "learning_rate": 0.00012885940727865334,
+ "loss": 1.8384013175964355,
+ "step": 2549
+ },
+ {
+ "epoch": 35.41811846689895,
+ "grad_norm": 0.05410664901137352,
+ "learning_rate": 0.00012864352343596247,
+ "loss": 1.8481371402740479,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43205574912892,
+ "grad_norm": 0.0576784573495388,
+ "learning_rate": 0.0001284277800691355,
+ "loss": 1.8333299160003662,
+ "step": 2551
+ },
+ {
+ "epoch": 35.44599303135889,
+ "grad_norm": 0.05358795076608658,
+ "learning_rate": 0.00012821217735199298,
+ "loss": 1.8201930522918701,
+ "step": 2552
+ },
+ {
+ "epoch": 35.45993031358885,
+ "grad_norm": 0.05687721073627472,
+ "learning_rate": 0.00012799671545824212,
+ "loss": 1.838738203048706,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47386759581882,
+ "grad_norm": 0.05585115775465965,
+ "learning_rate": 0.0001277813945614768,
+ "loss": 1.8358794450759888,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48780487804878,
+ "grad_norm": 0.057034578174352646,
+ "learning_rate": 0.00012756621483517682,
+ "loss": 1.8355228900909424,
+ "step": 2555
+ },
+ {
+ "epoch": 35.501742160278745,
+ "grad_norm": 0.05627685412764549,
+ "learning_rate": 0.00012735117645270905,
+ "loss": 1.8351898193359375,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51567944250871,
+ "grad_norm": 0.05724332481622696,
+ "learning_rate": 0.00012713627958732567,
+ "loss": 1.833801031112671,
+ "step": 2557
+ },
+ {
+ "epoch": 35.52961672473867,
+ "grad_norm": 0.05503929778933525,
+ "learning_rate": 0.00012692152441216539,
+ "loss": 1.8311514854431152,
+ "step": 2558
+ },
+ {
+ "epoch": 35.54355400696864,
+ "grad_norm": 0.057043809443712234,
+ "learning_rate": 0.0001267069111002525,
+ "loss": 1.8318891525268555,
+ "step": 2559
+ },
+ {
+ "epoch": 35.55749128919861,
+ "grad_norm": 0.05669008940458298,
+ "learning_rate": 0.00012649243982449718,
+ "loss": 1.8333967924118042,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57142857142857,
+ "grad_norm": 0.05707597732543945,
+ "learning_rate": 0.00012627811075769502,
+ "loss": 1.8416266441345215,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58536585365854,
+ "grad_norm": 0.055258579552173615,
+ "learning_rate": 0.00012606392407252687,
+ "loss": 1.8428752422332764,
+ "step": 2562
+ },
+ {
+ "epoch": 35.599303135888505,
+ "grad_norm": 0.05575427785515785,
+ "learning_rate": 0.00012584987994155943,
+ "loss": 1.8269420862197876,
+ "step": 2563
+ },
+ {
+ "epoch": 35.613240418118465,
+ "grad_norm": 0.05566711723804474,
+ "learning_rate": 0.00012563597853724388,
+ "loss": 1.8391327857971191,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62717770034843,
+ "grad_norm": 0.057135093957185745,
+ "learning_rate": 0.000125422220031917,
+ "loss": 1.8499560356140137,
+ "step": 2565
+ },
+ {
+ "epoch": 35.641114982578394,
+ "grad_norm": 0.05612964928150177,
+ "learning_rate": 0.00012520860459779972,
+ "loss": 1.8274946212768555,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65505226480836,
+ "grad_norm": 0.05745973810553551,
+ "learning_rate": 0.0001249951324069986,
+ "loss": 1.818924903869629,
+ "step": 2567
+ },
+ {
+ "epoch": 35.66898954703833,
+ "grad_norm": 0.05643341690301895,
+ "learning_rate": 0.0001247818036315042,
+ "loss": 1.8365602493286133,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68292682926829,
+ "grad_norm": 0.054355110973119736,
+ "learning_rate": 0.00012456861844319155,
+ "loss": 1.8275706768035889,
+ "step": 2569
+ },
+ {
+ "epoch": 35.69686411149826,
+ "grad_norm": 0.05827079713344574,
+ "learning_rate": 0.00012435557701382012,
+ "loss": 1.8421577215194702,
+ "step": 2570
+ },
+ {
+ "epoch": 35.710801393728225,
+ "grad_norm": 0.055892039090394974,
+ "learning_rate": 0.0001241426795150336,
+ "loss": 1.8322241306304932,
+ "step": 2571
+ },
+ {
+ "epoch": 35.724738675958186,
+ "grad_norm": 0.05828035622835159,
+ "learning_rate": 0.00012392992611835973,
+ "loss": 1.8444736003875732,
+ "step": 2572
+ },
+ {
+ "epoch": 35.73867595818815,
+ "grad_norm": 0.05669068545103073,
+ "learning_rate": 0.0001237173169952098,
+ "loss": 1.8492958545684814,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75261324041812,
+ "grad_norm": 0.05769318342208862,
+ "learning_rate": 0.00012350485231687954,
+ "loss": 1.8362290859222412,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76655052264808,
+ "grad_norm": 0.059012528508901596,
+ "learning_rate": 0.0001232925322545476,
+ "loss": 1.8452755212783813,
+ "step": 2575
+ },
+ {
+ "epoch": 35.78048780487805,
+ "grad_norm": 0.0577283576130867,
+ "learning_rate": 0.0001230803569792765,
+ "loss": 1.8381261825561523,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79442508710802,
+ "grad_norm": 0.05785193666815758,
+ "learning_rate": 0.0001228683266620121,
+ "loss": 1.8344440460205078,
+ "step": 2577
+ },
+ {
+ "epoch": 35.80836236933798,
+ "grad_norm": 0.057778336107730865,
+ "learning_rate": 0.00012265644147358326,
+ "loss": 1.8465049266815186,
+ "step": 2578
+ },
+ {
+ "epoch": 35.822299651567945,
+ "grad_norm": 0.05725527182221413,
+ "learning_rate": 0.00012244470158470226,
+ "loss": 1.8357681035995483,
+ "step": 2579
+ },
+ {
+ "epoch": 35.836236933797906,
+ "grad_norm": 0.05706310644745827,
+ "learning_rate": 0.00012223310716596387,
+ "loss": 1.8430027961730957,
+ "step": 2580
+ },
+ {
+ "epoch": 35.850174216027874,
+ "grad_norm": 0.05754264071583748,
+ "learning_rate": 0.00012202165838784601,
+ "loss": 1.8457331657409668,
+ "step": 2581
+ },
+ {
+ "epoch": 35.86411149825784,
+ "grad_norm": 0.05796574801206589,
+ "learning_rate": 0.00012181035542070913,
+ "loss": 1.8521018028259277,
+ "step": 2582
+ },
+ {
+ "epoch": 35.8780487804878,
+ "grad_norm": 0.05767996609210968,
+ "learning_rate": 0.00012159919843479617,
+ "loss": 1.837098240852356,
+ "step": 2583
+ },
+ {
+ "epoch": 35.89198606271777,
+ "grad_norm": 0.05741669982671738,
+ "learning_rate": 0.00012138818760023257,
+ "loss": 1.8438975811004639,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90592334494774,
+ "grad_norm": 0.05797840282320976,
+ "learning_rate": 0.00012117732308702592,
+ "loss": 1.8390898704528809,
+ "step": 2585
+ },
+ {
+ "epoch": 35.9198606271777,
+ "grad_norm": 0.057005491107702255,
+ "learning_rate": 0.00012096660506506605,
+ "loss": 1.853393316268921,
+ "step": 2586
+ },
+ {
+ "epoch": 35.933797909407666,
+ "grad_norm": 0.058777451515197754,
+ "learning_rate": 0.00012075603370412443,
+ "loss": 1.8450913429260254,
+ "step": 2587
+ },
+ {
+ "epoch": 35.94773519163763,
+ "grad_norm": 0.05810392275452614,
+ "learning_rate": 0.00012054560917385476,
+ "loss": 1.8371872901916504,
+ "step": 2588
+ },
+ {
+ "epoch": 35.961672473867594,
+ "grad_norm": 0.05875588208436966,
+ "learning_rate": 0.00012033533164379225,
+ "loss": 1.8611619472503662,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97560975609756,
+ "grad_norm": 0.05872174724936485,
+ "learning_rate": 0.0001201252012833537,
+ "loss": 1.8622443675994873,
+ "step": 2590
+ },
+ {
+ "epoch": 35.98954703832753,
+ "grad_norm": 0.059304554015398026,
+ "learning_rate": 0.00011991521826183747,
+ "loss": 1.8395988941192627,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.04685443267226219,
+ "learning_rate": 0.00011970538274842282,
+ "loss": 1.3844654560089111,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 0.644555926322937,
+ "eval_runtime": 40.5311,
+ "eval_samples_per_second": 60.25,
+ "eval_steps_per_second": 0.493,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01393728222997,
+ "grad_norm": 0.05542989447712898,
+ "learning_rate": 0.00011949569491217073,
+ "loss": 1.816087007522583,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02787456445993,
+ "grad_norm": 0.05269062519073486,
+ "learning_rate": 0.00011928615492202269,
+ "loss": 1.8240373134613037,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041811846689896,
+ "grad_norm": 0.05406419187784195,
+ "learning_rate": 0.0001190767629468014,
+ "loss": 1.8276968002319336,
+ "step": 2595
+ },
+ {
+ "epoch": 36.055749128919864,
+ "grad_norm": 0.052498709410429,
+ "learning_rate": 0.00011886751915521007,
+ "loss": 1.8275878429412842,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069686411149824,
+ "grad_norm": 0.05208680033683777,
+ "learning_rate": 0.00011865842371583278,
+ "loss": 1.8087416887283325,
+ "step": 2597
+ },
+ {
+ "epoch": 36.08362369337979,
+ "grad_norm": 0.05346959829330444,
+ "learning_rate": 0.00011844947679713396,
+ "loss": 1.8090479373931885,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09756097560975,
+ "grad_norm": 0.052402183413505554,
+ "learning_rate": 0.00011824067856745812,
+ "loss": 1.8265495300292969,
+ "step": 2599
+ },
+ {
+ "epoch": 36.11149825783972,
+ "grad_norm": 0.051748018711805344,
+ "learning_rate": 0.00011803202919503064,
+ "loss": 1.8173667192459106,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12543554006969,
+ "grad_norm": 0.05207061767578125,
+ "learning_rate": 0.00011782352884795615,
+ "loss": 1.821485996246338,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13937282229965,
+ "grad_norm": 0.05230815336108208,
+ "learning_rate": 0.00011761517769421983,
+ "loss": 1.804908037185669,
+ "step": 2602
+ },
+ {
+ "epoch": 36.153310104529616,
+ "grad_norm": 0.052593570202589035,
+ "learning_rate": 0.00011740697590168635,
+ "loss": 1.8154864311218262,
+ "step": 2603
+ },
+ {
+ "epoch": 36.167247386759584,
+ "grad_norm": 0.05259322747588158,
+ "learning_rate": 0.00011719892363810018,
+ "loss": 1.8238556385040283,
+ "step": 2604
+ },
+ {
+ "epoch": 36.181184668989545,
+ "grad_norm": 0.052652861922979355,
+ "learning_rate": 0.00011699102107108539,
+ "loss": 1.8147733211517334,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19512195121951,
+ "grad_norm": 0.0530865415930748,
+ "learning_rate": 0.00011678326836814507,
+ "loss": 1.8086535930633545,
+ "step": 2606
+ },
+ {
+ "epoch": 36.20905923344948,
+ "grad_norm": 0.053423281759023666,
+ "learning_rate": 0.00011657566569666198,
+ "loss": 1.8174736499786377,
+ "step": 2607
+ },
+ {
+ "epoch": 36.22299651567944,
+ "grad_norm": 0.0524580255150795,
+ "learning_rate": 0.00011636821322389777,
+ "loss": 1.827101230621338,
+ "step": 2608
+ },
+ {
+ "epoch": 36.23693379790941,
+ "grad_norm": 0.05376284569501877,
+ "learning_rate": 0.00011616091111699333,
+ "loss": 1.8410698175430298,
+ "step": 2609
+ },
+ {
+ "epoch": 36.250871080139376,
+ "grad_norm": 0.05277293920516968,
+ "learning_rate": 0.00011595375954296786,
+ "loss": 1.8262073993682861,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26480836236934,
+ "grad_norm": 0.05415027216076851,
+ "learning_rate": 0.00011574675866871997,
+ "loss": 1.8311805725097656,
+ "step": 2611
+ },
+ {
+ "epoch": 36.278745644599304,
+ "grad_norm": 0.05142449960112572,
+ "learning_rate": 0.00011553990866102651,
+ "loss": 1.816835880279541,
+ "step": 2612
+ },
+ {
+ "epoch": 36.292682926829265,
+ "grad_norm": 0.053121037781238556,
+ "learning_rate": 0.00011533320968654265,
+ "loss": 1.818075180053711,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30662020905923,
+ "grad_norm": 0.051999714225530624,
+ "learning_rate": 0.00011512666191180204,
+ "loss": 1.8257043361663818,
+ "step": 2614
+ },
+ {
+ "epoch": 36.3205574912892,
+ "grad_norm": 0.05257652327418327,
+ "learning_rate": 0.00011492026550321653,
+ "loss": 1.814011573791504,
+ "step": 2615
+ },
+ {
+ "epoch": 36.33449477351916,
+ "grad_norm": 0.0541016161441803,
+ "learning_rate": 0.00011471402062707607,
+ "loss": 1.826779842376709,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34843205574913,
+ "grad_norm": 0.05318204313516617,
+ "learning_rate": 0.00011450792744954827,
+ "loss": 1.8242204189300537,
+ "step": 2617
+ },
+ {
+ "epoch": 36.362369337979096,
+ "grad_norm": 0.05453426018357277,
+ "learning_rate": 0.00011430198613667861,
+ "loss": 1.822621464729309,
+ "step": 2618
+ },
+ {
+ "epoch": 36.37630662020906,
+ "grad_norm": 0.05376817658543587,
+ "learning_rate": 0.00011409619685439064,
+ "loss": 1.8253248929977417,
+ "step": 2619
+ },
+ {
+ "epoch": 36.390243902439025,
+ "grad_norm": 0.05483992025256157,
+ "learning_rate": 0.00011389055976848477,
+ "loss": 1.8317890167236328,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40418118466899,
+ "grad_norm": 0.05439801886677742,
+ "learning_rate": 0.00011368507504463914,
+ "loss": 1.835724949836731,
+ "step": 2621
+ },
+ {
+ "epoch": 36.41811846689895,
+ "grad_norm": 0.055256184190511703,
+ "learning_rate": 0.00011347974284840914,
+ "loss": 1.823319911956787,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43205574912892,
+ "grad_norm": 0.05287355184555054,
+ "learning_rate": 0.00011327456334522725,
+ "loss": 1.8103458881378174,
+ "step": 2623
+ },
+ {
+ "epoch": 36.44599303135889,
+ "grad_norm": 0.05546991154551506,
+ "learning_rate": 0.00011306953670040275,
+ "loss": 1.8309146165847778,
+ "step": 2624
+ },
+ {
+ "epoch": 36.45993031358885,
+ "grad_norm": 0.05322236195206642,
+ "learning_rate": 0.00011286466307912184,
+ "loss": 1.821721076965332,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47386759581882,
+ "grad_norm": 0.054857514798641205,
+ "learning_rate": 0.00011265994264644783,
+ "loss": 1.8251492977142334,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48780487804878,
+ "grad_norm": 0.053303562104701996,
+ "learning_rate": 0.00011245537556732001,
+ "loss": 1.8257558345794678,
+ "step": 2627
+ },
+ {
+ "epoch": 36.501742160278745,
+ "grad_norm": 0.05594215914607048,
+ "learning_rate": 0.00011225096200655442,
+ "loss": 1.835466742515564,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51567944250871,
+ "grad_norm": 0.05624304339289665,
+ "learning_rate": 0.00011204670212884342,
+ "loss": 1.8186707496643066,
+ "step": 2629
+ },
+ {
+ "epoch": 36.52961672473867,
+ "grad_norm": 0.055144522339105606,
+ "learning_rate": 0.00011184259609875557,
+ "loss": 1.8301169872283936,
+ "step": 2630
+ },
+ {
+ "epoch": 36.54355400696864,
+ "grad_norm": 0.0535074919462204,
+ "learning_rate": 0.00011163864408073522,
+ "loss": 1.8138389587402344,
+ "step": 2631
+ },
+ {
+ "epoch": 36.55749128919861,
+ "grad_norm": 0.05484950542449951,
+ "learning_rate": 0.00011143484623910293,
+ "loss": 1.8223397731781006,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57142857142857,
+ "grad_norm": 0.05473049357533455,
+ "learning_rate": 0.00011123120273805496,
+ "loss": 1.8218905925750732,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58536585365854,
+ "grad_norm": 0.05588974803686142,
+ "learning_rate": 0.0001110277137416632,
+ "loss": 1.825666904449463,
+ "step": 2634
+ },
+ {
+ "epoch": 36.599303135888505,
+ "grad_norm": 0.05313410982489586,
+ "learning_rate": 0.00011082437941387512,
+ "loss": 1.8338199853897095,
+ "step": 2635
+ },
+ {
+ "epoch": 36.613240418118465,
+ "grad_norm": 0.056002989411354065,
+ "learning_rate": 0.00011062119991851322,
+ "loss": 1.8219728469848633,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62717770034843,
+ "grad_norm": 0.05519993603229523,
+ "learning_rate": 0.000110418175419276,
+ "loss": 1.826596736907959,
+ "step": 2637
+ },
+ {
+ "epoch": 36.641114982578394,
+ "grad_norm": 0.055067840963602066,
+ "learning_rate": 0.00011021530607973626,
+ "loss": 1.8218865394592285,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65505226480836,
+ "grad_norm": 0.0548541434109211,
+ "learning_rate": 0.00011001259206334235,
+ "loss": 1.8283402919769287,
+ "step": 2639
+ },
+ {
+ "epoch": 36.66898954703833,
+ "grad_norm": 0.05756586790084839,
+ "learning_rate": 0.00010981003353341721,
+ "loss": 1.8259823322296143,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68292682926829,
+ "grad_norm": 0.05620390921831131,
+ "learning_rate": 0.00010960763065315864,
+ "loss": 1.8149542808532715,
+ "step": 2641
+ },
+ {
+ "epoch": 36.69686411149826,
+ "grad_norm": 0.05768077075481415,
+ "learning_rate": 0.00010940538358563906,
+ "loss": 1.8275048732757568,
+ "step": 2642
+ },
+ {
+ "epoch": 36.710801393728225,
+ "grad_norm": 0.05531959980726242,
+ "learning_rate": 0.00010920329249380509,
+ "loss": 1.8295093774795532,
+ "step": 2643
+ },
+ {
+ "epoch": 36.724738675958186,
+ "grad_norm": 0.057147517800331116,
+ "learning_rate": 0.00010900135754047799,
+ "loss": 1.8217309713363647,
+ "step": 2644
+ },
+ {
+ "epoch": 36.73867595818815,
+ "grad_norm": 0.05488502234220505,
+ "learning_rate": 0.00010879957888835307,
+ "loss": 1.8418645858764648,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75261324041812,
+ "grad_norm": 0.05747298523783684,
+ "learning_rate": 0.00010859795669999981,
+ "loss": 1.825081467628479,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76655052264808,
+ "grad_norm": 0.05621277540922165,
+ "learning_rate": 0.00010839649113786137,
+ "loss": 1.8208179473876953,
+ "step": 2647
+ },
+ {
+ "epoch": 36.78048780487805,
+ "grad_norm": 0.05534190684556961,
+ "learning_rate": 0.00010819518236425514,
+ "loss": 1.8269200325012207,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79442508710802,
+ "grad_norm": 0.0570702850818634,
+ "learning_rate": 0.00010799403054137197,
+ "loss": 1.8310284614562988,
+ "step": 2649
+ },
+ {
+ "epoch": 36.80836236933798,
+ "grad_norm": 0.055876970291137695,
+ "learning_rate": 0.00010779303583127609,
+ "loss": 1.831892967224121,
+ "step": 2650
+ },
+ {
+ "epoch": 36.822299651567945,
+ "grad_norm": 0.05680249258875847,
+ "learning_rate": 0.00010759219839590535,
+ "loss": 1.8280737400054932,
+ "step": 2651
+ },
+ {
+ "epoch": 36.836236933797906,
+ "grad_norm": 0.056308578699827194,
+ "learning_rate": 0.00010739151839707089,
+ "loss": 1.844076156616211,
+ "step": 2652
+ },
+ {
+ "epoch": 36.850174216027874,
+ "grad_norm": 0.05752384290099144,
+ "learning_rate": 0.00010719099599645706,
+ "loss": 1.8403749465942383,
+ "step": 2653
+ },
+ {
+ "epoch": 36.86411149825784,
+ "grad_norm": 0.058181654661893845,
+ "learning_rate": 0.0001069906313556208,
+ "loss": 1.8354277610778809,
+ "step": 2654
+ },
+ {
+ "epoch": 36.8780487804878,
+ "grad_norm": 0.05747664347290993,
+ "learning_rate": 0.00010679042463599263,
+ "loss": 1.8315486907958984,
+ "step": 2655
+ },
+ {
+ "epoch": 36.89198606271777,
+ "grad_norm": 0.057498663663864136,
+ "learning_rate": 0.00010659037599887545,
+ "loss": 1.8256535530090332,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90592334494774,
+ "grad_norm": 0.0565464086830616,
+ "learning_rate": 0.0001063904856054446,
+ "loss": 1.8265036344528198,
+ "step": 2657
+ },
+ {
+ "epoch": 36.9198606271777,
+ "grad_norm": 0.057560283690690994,
+ "learning_rate": 0.00010619075361674836,
+ "loss": 1.822776198387146,
+ "step": 2658
+ },
+ {
+ "epoch": 36.933797909407666,
+ "grad_norm": 0.05598663538694382,
+ "learning_rate": 0.0001059911801937071,
+ "loss": 1.8307170867919922,
+ "step": 2659
+ },
+ {
+ "epoch": 36.94773519163763,
+ "grad_norm": 0.0580892413854599,
+ "learning_rate": 0.00010579176549711365,
+ "loss": 1.831838846206665,
+ "step": 2660
+ },
+ {
+ "epoch": 36.961672473867594,
+ "grad_norm": 0.056303758174180984,
+ "learning_rate": 0.00010559250968763265,
+ "loss": 1.837568998336792,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97560975609756,
+ "grad_norm": 0.05781455338001251,
+ "learning_rate": 0.00010539341292580086,
+ "loss": 1.8347523212432861,
+ "step": 2662
+ },
+ {
+ "epoch": 36.98954703832753,
+ "grad_norm": 0.05536883324384689,
+ "learning_rate": 0.00010519447537202729,
+ "loss": 1.8196232318878174,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.046378862112760544,
+ "learning_rate": 0.00010499569718659201,
+ "loss": 1.3591325283050537,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 0.6459980607032776,
+ "eval_runtime": 40.631,
+ "eval_samples_per_second": 60.102,
+ "eval_steps_per_second": 0.492,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01393728222997,
+ "grad_norm": 0.050504088401794434,
+ "learning_rate": 0.00010479707852964713,
+ "loss": 1.8197872638702393,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02787456445993,
+ "grad_norm": 0.05133911967277527,
+ "learning_rate": 0.00010459861956121611,
+ "loss": 1.8092283010482788,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041811846689896,
+ "grad_norm": 0.051401812583208084,
+ "learning_rate": 0.00010440032044119383,
+ "loss": 1.7972350120544434,
+ "step": 2667
+ },
+ {
+ "epoch": 37.055749128919864,
+ "grad_norm": 0.050701867789030075,
+ "learning_rate": 0.00010420218132934614,
+ "loss": 1.8077106475830078,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069686411149824,
+ "grad_norm": 0.05219056457281113,
+ "learning_rate": 0.00010400420238531023,
+ "loss": 1.7995078563690186,
+ "step": 2669
+ },
+ {
+ "epoch": 37.08362369337979,
+ "grad_norm": 0.05003709718585014,
+ "learning_rate": 0.00010380638376859416,
+ "loss": 1.8166279792785645,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09756097560975,
+ "grad_norm": 0.051945704966783524,
+ "learning_rate": 0.00010360872563857682,
+ "loss": 1.8085274696350098,
+ "step": 2671
+ },
+ {
+ "epoch": 37.11149825783972,
+ "grad_norm": 0.05073661729693413,
+ "learning_rate": 0.0001034112281545079,
+ "loss": 1.794724941253662,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12543554006969,
+ "grad_norm": 0.051837258040905,
+ "learning_rate": 0.00010321389147550725,
+ "loss": 1.8101842403411865,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13937282229965,
+ "grad_norm": 0.051624711602926254,
+ "learning_rate": 0.00010301671576056588,
+ "loss": 1.8128712177276611,
+ "step": 2674
+ },
+ {
+ "epoch": 37.153310104529616,
+ "grad_norm": 0.05337493121623993,
+ "learning_rate": 0.00010281970116854436,
+ "loss": 1.8105556964874268,
+ "step": 2675
+ },
+ {
+ "epoch": 37.167247386759584,
+ "grad_norm": 0.05040125921368599,
+ "learning_rate": 0.00010262284785817392,
+ "loss": 1.801112413406372,
+ "step": 2676
+ },
+ {
+ "epoch": 37.181184668989545,
+ "grad_norm": 0.05222955718636513,
+ "learning_rate": 0.00010242615598805574,
+ "loss": 1.8220586776733398,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19512195121951,
+ "grad_norm": 0.052356619387865067,
+ "learning_rate": 0.00010222962571666088,
+ "loss": 1.8208140134811401,
+ "step": 2678
+ },
+ {
+ "epoch": 37.20905923344948,
+ "grad_norm": 0.05221114307641983,
+ "learning_rate": 0.00010203325720233032,
+ "loss": 1.8082315921783447,
+ "step": 2679
+ },
+ {
+ "epoch": 37.22299651567944,
+ "grad_norm": 0.05170206353068352,
+ "learning_rate": 0.00010183705060327433,
+ "loss": 1.8159395456314087,
+ "step": 2680
+ },
+ {
+ "epoch": 37.23693379790941,
+ "grad_norm": 0.05091318488121033,
+ "learning_rate": 0.00010164100607757346,
+ "loss": 1.8138477802276611,
+ "step": 2681
+ },
+ {
+ "epoch": 37.250871080139376,
+ "grad_norm": 0.05228526517748833,
+ "learning_rate": 0.00010144512378317687,
+ "loss": 1.8149440288543701,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26480836236934,
+ "grad_norm": 0.053375136107206345,
+ "learning_rate": 0.00010124940387790354,
+ "loss": 1.8169589042663574,
+ "step": 2683
+ },
+ {
+ "epoch": 37.278745644599304,
+ "grad_norm": 0.05215807631611824,
+ "learning_rate": 0.00010105384651944148,
+ "loss": 1.8025376796722412,
+ "step": 2684
+ },
+ {
+ "epoch": 37.292682926829265,
+ "grad_norm": 0.05166545882821083,
+ "learning_rate": 0.00010085845186534769,
+ "loss": 1.815457820892334,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30662020905923,
+ "grad_norm": 0.05314334109425545,
+ "learning_rate": 0.00010066322007304819,
+ "loss": 1.8092843294143677,
+ "step": 2686
+ },
+ {
+ "epoch": 37.3205574912892,
+ "grad_norm": 0.0525335855782032,
+ "learning_rate": 0.00010046815129983757,
+ "loss": 1.8068681955337524,
+ "step": 2687
+ },
+ {
+ "epoch": 37.33449477351916,
+ "grad_norm": 0.05154058337211609,
+ "learning_rate": 0.00010027324570287925,
+ "loss": 1.8220410346984863,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34843205574913,
+ "grad_norm": 0.05400070548057556,
+ "learning_rate": 0.00010007850343920519,
+ "loss": 1.8106049299240112,
+ "step": 2689
+ },
+ {
+ "epoch": 37.362369337979096,
+ "grad_norm": 0.051827941089868546,
+ "learning_rate": 9.988392466571572e-05,
+ "loss": 1.8139021396636963,
+ "step": 2690
+ },
+ {
+ "epoch": 37.37630662020906,
+ "grad_norm": 0.053685542196035385,
+ "learning_rate": 9.968950953917945e-05,
+ "loss": 1.8051722049713135,
+ "step": 2691
+ },
+ {
+ "epoch": 37.390243902439025,
+ "grad_norm": 0.05291858687996864,
+ "learning_rate": 9.949525821623309e-05,
+ "loss": 1.806849479675293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40418118466899,
+ "grad_norm": 0.05310729518532753,
+ "learning_rate": 9.930117085338156e-05,
+ "loss": 1.7940170764923096,
+ "step": 2693
+ },
+ {
+ "epoch": 37.41811846689895,
+ "grad_norm": 0.05288896709680557,
+ "learning_rate": 9.910724760699731e-05,
+ "loss": 1.824284553527832,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43205574912892,
+ "grad_norm": 0.052886299788951874,
+ "learning_rate": 9.891348863332093e-05,
+ "loss": 1.8206238746643066,
+ "step": 2695
+ },
+ {
+ "epoch": 37.44599303135889,
+ "grad_norm": 0.05356050282716751,
+ "learning_rate": 9.871989408846051e-05,
+ "loss": 1.8159115314483643,
+ "step": 2696
+ },
+ {
+ "epoch": 37.45993031358885,
+ "grad_norm": 0.053028833121061325,
+ "learning_rate": 9.852646412839178e-05,
+ "loss": 1.8148610591888428,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47386759581882,
+ "grad_norm": 0.05355606600642204,
+ "learning_rate": 9.833319890895756e-05,
+ "loss": 1.8167890310287476,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48780487804878,
+ "grad_norm": 0.053764909505844116,
+ "learning_rate": 9.814009858586817e-05,
+ "loss": 1.8095712661743164,
+ "step": 2699
+ },
+ {
+ "epoch": 37.501742160278745,
+ "grad_norm": 0.05224880203604698,
+ "learning_rate": 9.794716331470131e-05,
+ "loss": 1.8151801824569702,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51567944250871,
+ "grad_norm": 0.05442487820982933,
+ "learning_rate": 9.77543932509012e-05,
+ "loss": 1.8090888261795044,
+ "step": 2701
+ },
+ {
+ "epoch": 37.52961672473867,
+ "grad_norm": 0.05287255719304085,
+ "learning_rate": 9.756178854977925e-05,
+ "loss": 1.8190618753433228,
+ "step": 2702
+ },
+ {
+ "epoch": 37.54355400696864,
+ "grad_norm": 0.05432182550430298,
+ "learning_rate": 9.736934936651362e-05,
+ "loss": 1.8103792667388916,
+ "step": 2703
+ },
+ {
+ "epoch": 37.55749128919861,
+ "grad_norm": 0.05345293879508972,
+ "learning_rate": 9.717707585614916e-05,
+ "loss": 1.8107848167419434,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57142857142857,
+ "grad_norm": 0.05567241832613945,
+ "learning_rate": 9.698496817359698e-05,
+ "loss": 1.822261095046997,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58536585365854,
+ "grad_norm": 0.0534495934844017,
+ "learning_rate": 9.679302647363476e-05,
+ "loss": 1.8195551633834839,
+ "step": 2706
+ },
+ {
+ "epoch": 37.599303135888505,
+ "grad_norm": 0.05432790517807007,
+ "learning_rate": 9.660125091090675e-05,
+ "loss": 1.8119161128997803,
+ "step": 2707
+ },
+ {
+ "epoch": 37.613240418118465,
+ "grad_norm": 0.05465250089764595,
+ "learning_rate": 9.64096416399228e-05,
+ "loss": 1.8269178867340088,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62717770034843,
+ "grad_norm": 0.05540885776281357,
+ "learning_rate": 9.621819881505918e-05,
+ "loss": 1.8170043230056763,
+ "step": 2709
+ },
+ {
+ "epoch": 37.641114982578394,
+ "grad_norm": 0.053956300020217896,
+ "learning_rate": 9.602692259055767e-05,
+ "loss": 1.8141452074050903,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65505226480836,
+ "grad_norm": 0.0530884824693203,
+ "learning_rate": 9.583581312052646e-05,
+ "loss": 1.8228175640106201,
+ "step": 2711
+ },
+ {
+ "epoch": 37.66898954703833,
+ "grad_norm": 0.05414469167590141,
+ "learning_rate": 9.564487055893867e-05,
+ "loss": 1.8148912191390991,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68292682926829,
+ "grad_norm": 0.054135460406541824,
+ "learning_rate": 9.545409505963338e-05,
+ "loss": 1.8189573287963867,
+ "step": 2713
+ },
+ {
+ "epoch": 37.69686411149826,
+ "grad_norm": 0.05427670478820801,
+ "learning_rate": 9.526348677631499e-05,
+ "loss": 1.8352646827697754,
+ "step": 2714
+ },
+ {
+ "epoch": 37.710801393728225,
+ "grad_norm": 0.054425179958343506,
+ "learning_rate": 9.507304586255311e-05,
+ "loss": 1.8101688623428345,
+ "step": 2715
+ },
+ {
+ "epoch": 37.724738675958186,
+ "grad_norm": 0.05361952632665634,
+ "learning_rate": 9.488277247178267e-05,
+ "loss": 1.822049856185913,
+ "step": 2716
+ },
+ {
+ "epoch": 37.73867595818815,
+ "grad_norm": 0.053644467145204544,
+ "learning_rate": 9.469266675730319e-05,
+ "loss": 1.8222166299819946,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75261324041812,
+ "grad_norm": 0.054326146841049194,
+ "learning_rate": 9.450272887227983e-05,
+ "loss": 1.8122522830963135,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76655052264808,
+ "grad_norm": 0.05346079170703888,
+ "learning_rate": 9.431295896974182e-05,
+ "loss": 1.8157234191894531,
+ "step": 2719
+ },
+ {
+ "epoch": 37.78048780487805,
+ "grad_norm": 0.05502757802605629,
+ "learning_rate": 9.412335720258341e-05,
+ "loss": 1.8146154880523682,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79442508710802,
+ "grad_norm": 0.05353285372257233,
+ "learning_rate": 9.393392372356335e-05,
+ "loss": 1.81754732131958,
+ "step": 2721
+ },
+ {
+ "epoch": 37.80836236933798,
+ "grad_norm": 0.05462120845913887,
+ "learning_rate": 9.374465868530477e-05,
+ "loss": 1.8201050758361816,
+ "step": 2722
+ },
+ {
+ "epoch": 37.822299651567945,
+ "grad_norm": 0.05272475257515907,
+ "learning_rate": 9.355556224029515e-05,
+ "loss": 1.8173773288726807,
+ "step": 2723
+ },
+ {
+ "epoch": 37.836236933797906,
+ "grad_norm": 0.056003544479608536,
+ "learning_rate": 9.336663454088593e-05,
+ "loss": 1.8211115598678589,
+ "step": 2724
+ },
+ {
+ "epoch": 37.850174216027874,
+ "grad_norm": 0.05486152321100235,
+ "learning_rate": 9.317787573929282e-05,
+ "loss": 1.8221628665924072,
+ "step": 2725
+ },
+ {
+ "epoch": 37.86411149825784,
+ "grad_norm": 0.05549563467502594,
+ "learning_rate": 9.298928598759541e-05,
+ "loss": 1.8200361728668213,
+ "step": 2726
+ },
+ {
+ "epoch": 37.8780487804878,
+ "grad_norm": 0.05407272279262543,
+ "learning_rate": 9.280086543773698e-05,
+ "loss": 1.814354658126831,
+ "step": 2727
+ },
+ {
+ "epoch": 37.89198606271777,
+ "grad_norm": 0.05526091903448105,
+ "learning_rate": 9.26126142415247e-05,
+ "loss": 1.8025412559509277,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90592334494774,
+ "grad_norm": 0.05579185113310814,
+ "learning_rate": 9.2424532550629e-05,
+ "loss": 1.8137646913528442,
+ "step": 2729
+ },
+ {
+ "epoch": 37.9198606271777,
+ "grad_norm": 0.054667308926582336,
+ "learning_rate": 9.223662051658408e-05,
+ "loss": 1.8100636005401611,
+ "step": 2730
+ },
+ {
+ "epoch": 37.933797909407666,
+ "grad_norm": 0.054524678736925125,
+ "learning_rate": 9.204887829078709e-05,
+ "loss": 1.8058973550796509,
+ "step": 2731
+ },
+ {
+ "epoch": 37.94773519163763,
+ "grad_norm": 0.05454818904399872,
+ "learning_rate": 9.186130602449861e-05,
+ "loss": 1.8153016567230225,
+ "step": 2732
+ },
+ {
+ "epoch": 37.961672473867594,
+ "grad_norm": 0.05491559952497482,
+ "learning_rate": 9.167390386884224e-05,
+ "loss": 1.8142199516296387,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97560975609756,
+ "grad_norm": 0.055562835186719894,
+ "learning_rate": 9.148667197480455e-05,
+ "loss": 1.8082599639892578,
+ "step": 2734
+ },
+ {
+ "epoch": 37.98954703832753,
+ "grad_norm": 0.055928416550159454,
+ "learning_rate": 9.129961049323494e-05,
+ "loss": 1.8185434341430664,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.044970303773880005,
+ "learning_rate": 9.111271957484519e-05,
+ "loss": 1.346644401550293,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 0.6474234461784363,
+ "eval_runtime": 40.6464,
+ "eval_samples_per_second": 60.079,
+ "eval_steps_per_second": 0.492,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01393728222997,
+ "grad_norm": 0.05179993808269501,
+ "learning_rate": 9.092599937021031e-05,
+ "loss": 1.7899185419082642,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02787456445993,
+ "grad_norm": 0.04930533468723297,
+ "learning_rate": 9.073945002976715e-05,
+ "loss": 1.7904696464538574,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041811846689896,
+ "grad_norm": 0.052794501185417175,
+ "learning_rate": 9.055307170381522e-05,
+ "loss": 1.7964375019073486,
+ "step": 2739
+ },
+ {
+ "epoch": 38.055749128919864,
+ "grad_norm": 0.049202270805835724,
+ "learning_rate": 9.036686454251613e-05,
+ "loss": 1.790614128112793,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069686411149824,
+ "grad_norm": 0.04903653636574745,
+ "learning_rate": 9.018082869589369e-05,
+ "loss": 1.803176999092102,
+ "step": 2741
+ },
+ {
+ "epoch": 38.08362369337979,
+ "grad_norm": 0.04976186528801918,
+ "learning_rate": 8.999496431383368e-05,
+ "loss": 1.8021293878555298,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09756097560975,
+ "grad_norm": 0.05226270109415054,
+ "learning_rate": 8.980927154608341e-05,
+ "loss": 1.7997218370437622,
+ "step": 2743
+ },
+ {
+ "epoch": 38.11149825783972,
+ "grad_norm": 0.04967877268791199,
+ "learning_rate": 8.962375054225255e-05,
+ "loss": 1.8076910972595215,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12543554006969,
+ "grad_norm": 0.05012716352939606,
+ "learning_rate": 8.943840145181177e-05,
+ "loss": 1.7991023063659668,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13937282229965,
+ "grad_norm": 0.05108457803726196,
+ "learning_rate": 8.925322442409361e-05,
+ "loss": 1.8080018758773804,
+ "step": 2746
+ },
+ {
+ "epoch": 38.153310104529616,
+ "grad_norm": 0.05005490034818649,
+ "learning_rate": 8.906821960829163e-05,
+ "loss": 1.7945563793182373,
+ "step": 2747
+ },
+ {
+ "epoch": 38.167247386759584,
+ "grad_norm": 0.050113365054130554,
+ "learning_rate": 8.888338715346113e-05,
+ "loss": 1.802520990371704,
+ "step": 2748
+ },
+ {
+ "epoch": 38.181184668989545,
+ "grad_norm": 0.05096133425831795,
+ "learning_rate": 8.869872720851831e-05,
+ "loss": 1.7994227409362793,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19512195121951,
+ "grad_norm": 0.05119607597589493,
+ "learning_rate": 8.851423992224012e-05,
+ "loss": 1.8115901947021484,
+ "step": 2750
+ },
+ {
+ "epoch": 38.20905923344948,
+ "grad_norm": 0.051627811044454575,
+ "learning_rate": 8.832992544326473e-05,
+ "loss": 1.800724744796753,
+ "step": 2751
+ },
+ {
+ "epoch": 38.22299651567944,
+ "grad_norm": 0.05053049698472023,
+ "learning_rate": 8.814578392009104e-05,
+ "loss": 1.7860815525054932,
+ "step": 2752
+ },
+ {
+ "epoch": 38.23693379790941,
+ "grad_norm": 0.05122068524360657,
+ "learning_rate": 8.796181550107857e-05,
+ "loss": 1.7976000308990479,
+ "step": 2753
+ },
+ {
+ "epoch": 38.250871080139376,
+ "grad_norm": 0.05197495222091675,
+ "learning_rate": 8.777802033444712e-05,
+ "loss": 1.811307430267334,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26480836236934,
+ "grad_norm": 0.051113326102495193,
+ "learning_rate": 8.75943985682774e-05,
+ "loss": 1.7974506616592407,
+ "step": 2755
+ },
+ {
+ "epoch": 38.278745644599304,
+ "grad_norm": 0.05044601857662201,
+ "learning_rate": 8.741095035051017e-05,
+ "loss": 1.8004801273345947,
+ "step": 2756
+ },
+ {
+ "epoch": 38.292682926829265,
+ "grad_norm": 0.05080671235918999,
+ "learning_rate": 8.722767582894613e-05,
+ "loss": 1.7960381507873535,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30662020905923,
+ "grad_norm": 0.05112270265817642,
+ "learning_rate": 8.704457515124636e-05,
+ "loss": 1.8070579767227173,
+ "step": 2758
+ },
+ {
+ "epoch": 38.3205574912892,
+ "grad_norm": 0.050059717148542404,
+ "learning_rate": 8.686164846493176e-05,
+ "loss": 1.7989699840545654,
+ "step": 2759
+ },
+ {
+ "epoch": 38.33449477351916,
+ "grad_norm": 0.05151006579399109,
+ "learning_rate": 8.667889591738317e-05,
+ "loss": 1.793050765991211,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34843205574913,
+ "grad_norm": 0.05045556277036667,
+ "learning_rate": 8.649631765584083e-05,
+ "loss": 1.7945470809936523,
+ "step": 2761
+ },
+ {
+ "epoch": 38.362369337979096,
+ "grad_norm": 0.05152078717947006,
+ "learning_rate": 8.631391382740482e-05,
+ "loss": 1.8155558109283447,
+ "step": 2762
+ },
+ {
+ "epoch": 38.37630662020906,
+ "grad_norm": 0.049696408212184906,
+ "learning_rate": 8.613168457903459e-05,
+ "loss": 1.7929422855377197,
+ "step": 2763
+ },
+ {
+ "epoch": 38.390243902439025,
+ "grad_norm": 0.050947509706020355,
+ "learning_rate": 8.594963005754901e-05,
+ "loss": 1.8029417991638184,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40418118466899,
+ "grad_norm": 0.04988504573702812,
+ "learning_rate": 8.57677504096261e-05,
+ "loss": 1.807220697402954,
+ "step": 2765
+ },
+ {
+ "epoch": 38.41811846689895,
+ "grad_norm": 0.05031213164329529,
+ "learning_rate": 8.558604578180301e-05,
+ "loss": 1.820298433303833,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43205574912892,
+ "grad_norm": 0.05067938566207886,
+ "learning_rate": 8.540451632047593e-05,
+ "loss": 1.8025063276290894,
+ "step": 2767
+ },
+ {
+ "epoch": 38.44599303135889,
+ "grad_norm": 0.050599709153175354,
+ "learning_rate": 8.522316217189972e-05,
+ "loss": 1.7912819385528564,
+ "step": 2768
+ },
+ {
+ "epoch": 38.45993031358885,
+ "grad_norm": 0.05000261962413788,
+ "learning_rate": 8.504198348218821e-05,
+ "loss": 1.8116015195846558,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47386759581882,
+ "grad_norm": 0.05164877697825432,
+ "learning_rate": 8.486098039731384e-05,
+ "loss": 1.7894442081451416,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48780487804878,
+ "grad_norm": 0.05037420988082886,
+ "learning_rate": 8.46801530631075e-05,
+ "loss": 1.8002345561981201,
+ "step": 2771
+ },
+ {
+ "epoch": 38.501742160278745,
+ "grad_norm": 0.05061398074030876,
+ "learning_rate": 8.449950162525859e-05,
+ "loss": 1.8103997707366943,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51567944250871,
+ "grad_norm": 0.05141839757561684,
+ "learning_rate": 8.431902622931443e-05,
+ "loss": 1.7920000553131104,
+ "step": 2773
+ },
+ {
+ "epoch": 38.52961672473867,
+ "grad_norm": 0.050547219812870026,
+ "learning_rate": 8.413872702068119e-05,
+ "loss": 1.8011581897735596,
+ "step": 2774
+ },
+ {
+ "epoch": 38.54355400696864,
+ "grad_norm": 0.052665967494249344,
+ "learning_rate": 8.395860414462238e-05,
+ "loss": 1.8171170949935913,
+ "step": 2775
+ },
+ {
+ "epoch": 38.55749128919861,
+ "grad_norm": 0.05148577317595482,
+ "learning_rate": 8.377865774625985e-05,
+ "loss": 1.7976020574569702,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57142857142857,
+ "grad_norm": 0.05378367006778717,
+ "learning_rate": 8.359888797057316e-05,
+ "loss": 1.8061752319335938,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58536585365854,
+ "grad_norm": 0.05071459710597992,
+ "learning_rate": 8.341929496239958e-05,
+ "loss": 1.8084588050842285,
+ "step": 2778
+ },
+ {
+ "epoch": 38.599303135888505,
+ "grad_norm": 0.05229831859469414,
+ "learning_rate": 8.323987886643402e-05,
+ "loss": 1.8100000619888306,
+ "step": 2779
+ },
+ {
+ "epoch": 38.613240418118465,
+ "grad_norm": 0.052434664219617844,
+ "learning_rate": 8.306063982722855e-05,
+ "loss": 1.8119068145751953,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62717770034843,
+ "grad_norm": 0.05269220098853111,
+ "learning_rate": 8.288157798919318e-05,
+ "loss": 1.799889087677002,
+ "step": 2781
+ },
+ {
+ "epoch": 38.641114982578394,
+ "grad_norm": 0.05271878093481064,
+ "learning_rate": 8.27026934965945e-05,
+ "loss": 1.792294979095459,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65505226480836,
+ "grad_norm": 0.05291058123111725,
+ "learning_rate": 8.25239864935566e-05,
+ "loss": 1.7999838590621948,
+ "step": 2783
+ },
+ {
+ "epoch": 38.66898954703833,
+ "grad_norm": 0.05346683785319328,
+ "learning_rate": 8.23454571240605e-05,
+ "loss": 1.801847219467163,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68292682926829,
+ "grad_norm": 0.053118057548999786,
+ "learning_rate": 8.216710553194416e-05,
+ "loss": 1.7951825857162476,
+ "step": 2785
+ },
+ {
+ "epoch": 38.69686411149826,
+ "grad_norm": 0.0527576319873333,
+ "learning_rate": 8.198893186090222e-05,
+ "loss": 1.8107949495315552,
+ "step": 2786
+ },
+ {
+ "epoch": 38.710801393728225,
+ "grad_norm": 0.05188573896884918,
+ "learning_rate": 8.181093625448585e-05,
+ "loss": 1.8066956996917725,
+ "step": 2787
+ },
+ {
+ "epoch": 38.724738675958186,
+ "grad_norm": 0.05341636762022972,
+ "learning_rate": 8.163311885610299e-05,
+ "loss": 1.8036713600158691,
+ "step": 2788
+ },
+ {
+ "epoch": 38.73867595818815,
+ "grad_norm": 0.051814883947372437,
+ "learning_rate": 8.145547980901791e-05,
+ "loss": 1.8106606006622314,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75261324041812,
+ "grad_norm": 0.05374172329902649,
+ "learning_rate": 8.127801925635126e-05,
+ "loss": 1.7904407978057861,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76655052264808,
+ "grad_norm": 0.052160196006298065,
+ "learning_rate": 8.110073734107954e-05,
+ "loss": 1.797128677368164,
+ "step": 2791
+ },
+ {
+ "epoch": 38.78048780487805,
+ "grad_norm": 0.052794717252254486,
+ "learning_rate": 8.092363420603584e-05,
+ "loss": 1.8085696697235107,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79442508710802,
+ "grad_norm": 0.05286112427711487,
+ "learning_rate": 8.074670999390896e-05,
+ "loss": 1.8027613162994385,
+ "step": 2793
+ },
+ {
+ "epoch": 38.80836236933798,
+ "grad_norm": 0.052660733461380005,
+ "learning_rate": 8.056996484724342e-05,
+ "loss": 1.8135063648223877,
+ "step": 2794
+ },
+ {
+ "epoch": 38.822299651567945,
+ "grad_norm": 0.052824389189481735,
+ "learning_rate": 8.039339890843958e-05,
+ "loss": 1.7999730110168457,
+ "step": 2795
+ },
+ {
+ "epoch": 38.836236933797906,
+ "grad_norm": 0.05390750244259834,
+ "learning_rate": 8.02170123197535e-05,
+ "loss": 1.7879369258880615,
+ "step": 2796
+ },
+ {
+ "epoch": 38.850174216027874,
+ "grad_norm": 0.053744010627269745,
+ "learning_rate": 8.004080522329674e-05,
+ "loss": 1.811363935470581,
+ "step": 2797
+ },
+ {
+ "epoch": 38.86411149825784,
+ "grad_norm": 0.05401702970266342,
+ "learning_rate": 7.986477776103601e-05,
+ "loss": 1.8049407005310059,
+ "step": 2798
+ },
+ {
+ "epoch": 38.8780487804878,
+ "grad_norm": 0.05376699939370155,
+ "learning_rate": 7.968893007479343e-05,
+ "loss": 1.8111652135849,
+ "step": 2799
+ },
+ {
+ "epoch": 38.89198606271777,
+ "grad_norm": 0.05381734296679497,
+ "learning_rate": 7.951326230624658e-05,
+ "loss": 1.8312594890594482,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90592334494774,
+ "grad_norm": 0.05387615039944649,
+ "learning_rate": 7.933777459692756e-05,
+ "loss": 1.805098295211792,
+ "step": 2801
+ },
+ {
+ "epoch": 38.9198606271777,
+ "grad_norm": 0.05285770446062088,
+ "learning_rate": 7.916246708822373e-05,
+ "loss": 1.8160698413848877,
+ "step": 2802
+ },
+ {
+ "epoch": 38.933797909407666,
+ "grad_norm": 0.05545372888445854,
+ "learning_rate": 7.898733992137715e-05,
+ "loss": 1.8070565462112427,
+ "step": 2803
+ },
+ {
+ "epoch": 38.94773519163763,
+ "grad_norm": 0.05404449254274368,
+ "learning_rate": 7.881239323748475e-05,
+ "loss": 1.8055229187011719,
+ "step": 2804
+ },
+ {
+ "epoch": 38.961672473867594,
+ "grad_norm": 0.05500908941030502,
+ "learning_rate": 7.863762717749771e-05,
+ "loss": 1.798984169960022,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97560975609756,
+ "grad_norm": 0.05408451333642006,
+ "learning_rate": 7.846304188222189e-05,
+ "loss": 1.8082829713821411,
+ "step": 2806
+ },
+ {
+ "epoch": 38.98954703832753,
+ "grad_norm": 0.0534386970102787,
+ "learning_rate": 7.828863749231777e-05,
+ "loss": 1.8156507015228271,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.044039830565452576,
+ "learning_rate": 7.811441414829958e-05,
+ "loss": 1.3510627746582031,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 0.6490727663040161,
+ "eval_runtime": 41.0373,
+ "eval_samples_per_second": 59.507,
+ "eval_steps_per_second": 0.487,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01393728222997,
+ "grad_norm": 0.0485052689909935,
+ "learning_rate": 7.794037199053598e-05,
+ "loss": 1.787034511566162,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02787456445993,
+ "grad_norm": 0.046773314476013184,
+ "learning_rate": 7.776651115924959e-05,
+ "loss": 1.7824962139129639,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041811846689896,
+ "grad_norm": 0.04776127636432648,
+ "learning_rate": 7.759283179451704e-05,
+ "loss": 1.784999132156372,
+ "step": 2811
+ },
+ {
+ "epoch": 39.055749128919864,
+ "grad_norm": 0.04702889546751976,
+ "learning_rate": 7.741933403626848e-05,
+ "loss": 1.7887163162231445,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069686411149824,
+ "grad_norm": 0.0472981221973896,
+ "learning_rate": 7.724601802428805e-05,
+ "loss": 1.7778058052062988,
+ "step": 2813
+ },
+ {
+ "epoch": 39.08362369337979,
+ "grad_norm": 0.04829077050089836,
+ "learning_rate": 7.707288389821329e-05,
+ "loss": 1.8164949417114258,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09756097560975,
+ "grad_norm": 0.04883906990289688,
+ "learning_rate": 7.689993179753519e-05,
+ "loss": 1.787117838859558,
+ "step": 2815
+ },
+ {
+ "epoch": 39.11149825783972,
+ "grad_norm": 0.0464581660926342,
+ "learning_rate": 7.672716186159829e-05,
+ "loss": 1.7883102893829346,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12543554006969,
+ "grad_norm": 0.047438278794288635,
+ "learning_rate": 7.655457422959993e-05,
+ "loss": 1.7923827171325684,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13937282229965,
+ "grad_norm": 0.04703788831830025,
+ "learning_rate": 7.638216904059122e-05,
+ "loss": 1.78762686252594,
+ "step": 2818
+ },
+ {
+ "epoch": 39.153310104529616,
+ "grad_norm": 0.049375683069229126,
+ "learning_rate": 7.620994643347559e-05,
+ "loss": 1.7970117330551147,
+ "step": 2819
+ },
+ {
+ "epoch": 39.167247386759584,
+ "grad_norm": 0.04796822741627693,
+ "learning_rate": 7.60379065470098e-05,
+ "loss": 1.778883457183838,
+ "step": 2820
+ },
+ {
+ "epoch": 39.181184668989545,
+ "grad_norm": 0.04719251021742821,
+ "learning_rate": 7.586604951980326e-05,
+ "loss": 1.7749544382095337,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19512195121951,
+ "grad_norm": 0.04893529415130615,
+ "learning_rate": 7.569437549031813e-05,
+ "loss": 1.787841796875,
+ "step": 2822
+ },
+ {
+ "epoch": 39.20905923344948,
+ "grad_norm": 0.04869271442294121,
+ "learning_rate": 7.55228845968691e-05,
+ "loss": 1.8007760047912598,
+ "step": 2823
+ },
+ {
+ "epoch": 39.22299651567944,
+ "grad_norm": 0.04950389638543129,
+ "learning_rate": 7.535157697762313e-05,
+ "loss": 1.7905025482177734,
+ "step": 2824
+ },
+ {
+ "epoch": 39.23693379790941,
+ "grad_norm": 0.0489855520427227,
+ "learning_rate": 7.518045277059979e-05,
+ "loss": 1.7977821826934814,
+ "step": 2825
+ },
+ {
+ "epoch": 39.250871080139376,
+ "grad_norm": 0.04881766065955162,
+ "learning_rate": 7.500951211367068e-05,
+ "loss": 1.7955210208892822,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26480836236934,
+ "grad_norm": 0.04834857955574989,
+ "learning_rate": 7.483875514455979e-05,
+ "loss": 1.7897529602050781,
+ "step": 2827
+ },
+ {
+ "epoch": 39.278745644599304,
+ "grad_norm": 0.04898613691329956,
+ "learning_rate": 7.466818200084264e-05,
+ "loss": 1.7888939380645752,
+ "step": 2828
+ },
+ {
+ "epoch": 39.292682926829265,
+ "grad_norm": 0.04936131089925766,
+ "learning_rate": 7.449779281994712e-05,
+ "loss": 1.8096022605895996,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30662020905923,
+ "grad_norm": 0.04834993928670883,
+ "learning_rate": 7.43275877391528e-05,
+ "loss": 1.800262689590454,
+ "step": 2830
+ },
+ {
+ "epoch": 39.3205574912892,
+ "grad_norm": 0.049713220447301865,
+ "learning_rate": 7.415756689559061e-05,
+ "loss": 1.7950725555419922,
+ "step": 2831
+ },
+ {
+ "epoch": 39.33449477351916,
+ "grad_norm": 0.049898866564035416,
+ "learning_rate": 7.398773042624341e-05,
+ "loss": 1.7812716960906982,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34843205574913,
+ "grad_norm": 0.049175627529621124,
+ "learning_rate": 7.381807846794532e-05,
+ "loss": 1.7746937274932861,
+ "step": 2833
+ },
+ {
+ "epoch": 39.362369337979096,
+ "grad_norm": 0.04872290790081024,
+ "learning_rate": 7.3648611157382e-05,
+ "loss": 1.7824972867965698,
+ "step": 2834
+ },
+ {
+ "epoch": 39.37630662020906,
+ "grad_norm": 0.049543045461177826,
+ "learning_rate": 7.34793286310899e-05,
+ "loss": 1.793465256690979,
+ "step": 2835
+ },
+ {
+ "epoch": 39.390243902439025,
+ "grad_norm": 0.049746304750442505,
+ "learning_rate": 7.331023102545716e-05,
+ "loss": 1.792328119277954,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40418118466899,
+ "grad_norm": 0.04971722885966301,
+ "learning_rate": 7.314131847672268e-05,
+ "loss": 1.7932684421539307,
+ "step": 2837
+ },
+ {
+ "epoch": 39.41811846689895,
+ "grad_norm": 0.05128713324666023,
+ "learning_rate": 7.297259112097608e-05,
+ "loss": 1.8031436204910278,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43205574912892,
+ "grad_norm": 0.04928988218307495,
+ "learning_rate": 7.280404909415801e-05,
+ "loss": 1.7928621768951416,
+ "step": 2839
+ },
+ {
+ "epoch": 39.44599303135889,
+ "grad_norm": 0.05211673676967621,
+ "learning_rate": 7.263569253205973e-05,
+ "loss": 1.7749731540679932,
+ "step": 2840
+ },
+ {
+ "epoch": 39.45993031358885,
+ "grad_norm": 0.05039450526237488,
+ "learning_rate": 7.24675215703232e-05,
+ "loss": 1.7910631895065308,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47386759581882,
+ "grad_norm": 0.05047272890806198,
+ "learning_rate": 7.229953634444055e-05,
+ "loss": 1.8027098178863525,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48780487804878,
+ "grad_norm": 0.05042344331741333,
+ "learning_rate": 7.213173698975444e-05,
+ "loss": 1.8134506940841675,
+ "step": 2843
+ },
+ {
+ "epoch": 39.501742160278745,
+ "grad_norm": 0.0506373755633831,
+ "learning_rate": 7.1964123641458e-05,
+ "loss": 1.7785459756851196,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51567944250871,
+ "grad_norm": 0.04902270808815956,
+ "learning_rate": 7.179669643459405e-05,
+ "loss": 1.7985787391662598,
+ "step": 2845
+ },
+ {
+ "epoch": 39.52961672473867,
+ "grad_norm": 0.05162668600678444,
+ "learning_rate": 7.162945550405575e-05,
+ "loss": 1.7813456058502197,
+ "step": 2846
+ },
+ {
+ "epoch": 39.54355400696864,
+ "grad_norm": 0.05005355924367905,
+ "learning_rate": 7.146240098458613e-05,
+ "loss": 1.791803240776062,
+ "step": 2847
+ },
+ {
+ "epoch": 39.55749128919861,
+ "grad_norm": 0.050947558134794235,
+ "learning_rate": 7.129553301077801e-05,
+ "loss": 1.7949843406677246,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57142857142857,
+ "grad_norm": 0.05130480229854584,
+ "learning_rate": 7.112885171707378e-05,
+ "loss": 1.7962243556976318,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58536585365854,
+ "grad_norm": 0.049898259341716766,
+ "learning_rate": 7.096235723776562e-05,
+ "loss": 1.7940798997879028,
+ "step": 2850
+ },
+ {
+ "epoch": 39.599303135888505,
+ "grad_norm": 0.050370268523693085,
+ "learning_rate": 7.079604970699514e-05,
+ "loss": 1.8001989126205444,
+ "step": 2851
+ },
+ {
+ "epoch": 39.613240418118465,
+ "grad_norm": 0.05121270939707756,
+ "learning_rate": 7.062992925875332e-05,
+ "loss": 1.79237961769104,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62717770034843,
+ "grad_norm": 0.05014092102646828,
+ "learning_rate": 7.046399602688045e-05,
+ "loss": 1.804434895515442,
+ "step": 2853
+ },
+ {
+ "epoch": 39.641114982578394,
+ "grad_norm": 0.050814609974622726,
+ "learning_rate": 7.029825014506572e-05,
+ "loss": 1.7997581958770752,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65505226480836,
+ "grad_norm": 0.05023293197154999,
+ "learning_rate": 7.013269174684795e-05,
+ "loss": 1.8026623725891113,
+ "step": 2855
+ },
+ {
+ "epoch": 39.66898954703833,
+ "grad_norm": 0.051415517926216125,
+ "learning_rate": 6.996732096561422e-05,
+ "loss": 1.8061354160308838,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68292682926829,
+ "grad_norm": 0.04993635416030884,
+ "learning_rate": 6.980213793460092e-05,
+ "loss": 1.7835321426391602,
+ "step": 2857
+ },
+ {
+ "epoch": 39.69686411149826,
+ "grad_norm": 0.05157633125782013,
+ "learning_rate": 6.963714278689304e-05,
+ "loss": 1.8007519245147705,
+ "step": 2858
+ },
+ {
+ "epoch": 39.710801393728225,
+ "grad_norm": 0.050889477133750916,
+ "learning_rate": 6.947233565542417e-05,
+ "loss": 1.7806460857391357,
+ "step": 2859
+ },
+ {
+ "epoch": 39.724738675958186,
+ "grad_norm": 0.050646744668483734,
+ "learning_rate": 6.930771667297653e-05,
+ "loss": 1.780684471130371,
+ "step": 2860
+ },
+ {
+ "epoch": 39.73867595818815,
+ "grad_norm": 0.05104859918355942,
+ "learning_rate": 6.914328597218042e-05,
+ "loss": 1.796810507774353,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75261324041812,
+ "grad_norm": 0.051958777010440826,
+ "learning_rate": 6.8979043685515e-05,
+ "loss": 1.8104615211486816,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76655052264808,
+ "grad_norm": 0.05137504264712334,
+ "learning_rate": 6.881498994530708e-05,
+ "loss": 1.7988557815551758,
+ "step": 2863
+ },
+ {
+ "epoch": 39.78048780487805,
+ "grad_norm": 0.051518820226192474,
+ "learning_rate": 6.865112488373186e-05,
+ "loss": 1.7865259647369385,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79442508710802,
+ "grad_norm": 0.051753368228673935,
+ "learning_rate": 6.84874486328125e-05,
+ "loss": 1.8039045333862305,
+ "step": 2865
+ },
+ {
+ "epoch": 39.80836236933798,
+ "grad_norm": 0.0514046885073185,
+ "learning_rate": 6.832396132441993e-05,
+ "loss": 1.8040326833724976,
+ "step": 2866
+ },
+ {
+ "epoch": 39.822299651567945,
+ "grad_norm": 0.05259102210402489,
+ "learning_rate": 6.816066309027309e-05,
+ "loss": 1.8034707307815552,
+ "step": 2867
+ },
+ {
+ "epoch": 39.836236933797906,
+ "grad_norm": 0.05125397443771362,
+ "learning_rate": 6.799755406193815e-05,
+ "loss": 1.7860667705535889,
+ "step": 2868
+ },
+ {
+ "epoch": 39.850174216027874,
+ "grad_norm": 0.05120245739817619,
+ "learning_rate": 6.783463437082929e-05,
+ "loss": 1.7920024394989014,
+ "step": 2869
+ },
+ {
+ "epoch": 39.86411149825784,
+ "grad_norm": 0.05269347131252289,
+ "learning_rate": 6.767190414820792e-05,
+ "loss": 1.7958977222442627,
+ "step": 2870
+ },
+ {
+ "epoch": 39.8780487804878,
+ "grad_norm": 0.050947729498147964,
+ "learning_rate": 6.750936352518284e-05,
+ "loss": 1.7864038944244385,
+ "step": 2871
+ },
+ {
+ "epoch": 39.89198606271777,
+ "grad_norm": 0.052531689405441284,
+ "learning_rate": 6.734701263271011e-05,
+ "loss": 1.795427918434143,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90592334494774,
+ "grad_norm": 0.052431896328926086,
+ "learning_rate": 6.718485160159289e-05,
+ "loss": 1.7928128242492676,
+ "step": 2873
+ },
+ {
+ "epoch": 39.9198606271777,
+ "grad_norm": 0.053628548979759216,
+ "learning_rate": 6.702288056248145e-05,
+ "loss": 1.7986962795257568,
+ "step": 2874
+ },
+ {
+ "epoch": 39.933797909407666,
+ "grad_norm": 0.05186983942985535,
+ "learning_rate": 6.686109964587285e-05,
+ "loss": 1.7967369556427002,
+ "step": 2875
+ },
+ {
+ "epoch": 39.94773519163763,
+ "grad_norm": 0.05122850835323334,
+ "learning_rate": 6.669950898211106e-05,
+ "loss": 1.7836300134658813,
+ "step": 2876
+ },
+ {
+ "epoch": 39.961672473867594,
+ "grad_norm": 0.05265440791845322,
+ "learning_rate": 6.653810870138676e-05,
+ "loss": 1.801450490951538,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97560975609756,
+ "grad_norm": 0.053320810198783875,
+ "learning_rate": 6.637689893373729e-05,
+ "loss": 1.792264461517334,
+ "step": 2878
+ },
+ {
+ "epoch": 39.98954703832753,
+ "grad_norm": 0.05239804461598396,
+ "learning_rate": 6.621587980904651e-05,
+ "loss": 1.8015801906585693,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.04351874440908432,
+ "learning_rate": 6.605505145704437e-05,
+ "loss": 1.3430795669555664,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 0.6505019068717957,
+ "eval_runtime": 40.2814,
+ "eval_samples_per_second": 60.623,
+ "eval_steps_per_second": 0.497,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01393728222997,
+ "grad_norm": 0.04842356964945793,
+ "learning_rate": 6.58944140073077e-05,
+ "loss": 1.7894266843795776,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02787456445993,
+ "grad_norm": 0.04759506508708,
+ "learning_rate": 6.573396758925891e-05,
+ "loss": 1.7896207571029663,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041811846689896,
+ "grad_norm": 0.04793368652462959,
+ "learning_rate": 6.557371233216692e-05,
+ "loss": 1.7770864963531494,
+ "step": 2883
+ },
+ {
+ "epoch": 40.055749128919864,
+ "grad_norm": 0.04775947704911232,
+ "learning_rate": 6.541364836514646e-05,
+ "loss": 1.7689225673675537,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069686411149824,
+ "grad_norm": 0.04841279238462448,
+ "learning_rate": 6.525377581715829e-05,
+ "loss": 1.790536880493164,
+ "step": 2885
+ },
+ {
+ "epoch": 40.08362369337979,
+ "grad_norm": 0.04589880257844925,
+ "learning_rate": 6.509409481700864e-05,
+ "loss": 1.7804309129714966,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09756097560975,
+ "grad_norm": 0.04910781979560852,
+ "learning_rate": 6.49346054933496e-05,
+ "loss": 1.7737990617752075,
+ "step": 2887
+ },
+ {
+ "epoch": 40.11149825783972,
+ "grad_norm": 0.04736403003334999,
+ "learning_rate": 6.477530797467911e-05,
+ "loss": 1.7843401432037354,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12543554006969,
+ "grad_norm": 0.047884710133075714,
+ "learning_rate": 6.461620238934006e-05,
+ "loss": 1.7776315212249756,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13937282229965,
+ "grad_norm": 0.047823768109083176,
+ "learning_rate": 6.445728886552109e-05,
+ "loss": 1.7916157245635986,
+ "step": 2890
+ },
+ {
+ "epoch": 40.153310104529616,
+ "grad_norm": 0.04811791703104973,
+ "learning_rate": 6.429856753125573e-05,
+ "loss": 1.8077316284179688,
+ "step": 2891
+ },
+ {
+ "epoch": 40.167247386759584,
+ "grad_norm": 0.05039852485060692,
+ "learning_rate": 6.414003851442318e-05,
+ "loss": 1.7821904420852661,
+ "step": 2892
+ },
+ {
+ "epoch": 40.181184668989545,
+ "grad_norm": 0.04656902328133583,
+ "learning_rate": 6.398170194274722e-05,
+ "loss": 1.775067687034607,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19512195121951,
+ "grad_norm": 0.04947083443403244,
+ "learning_rate": 6.38235579437968e-05,
+ "loss": 1.7826871871948242,
+ "step": 2894
+ },
+ {
+ "epoch": 40.20905923344948,
+ "grad_norm": 0.04797012731432915,
+ "learning_rate": 6.366560664498572e-05,
+ "loss": 1.792678713798523,
+ "step": 2895
+ },
+ {
+ "epoch": 40.22299651567944,
+ "grad_norm": 0.048327650874853134,
+ "learning_rate": 6.350784817357242e-05,
+ "loss": 1.7605221271514893,
+ "step": 2896
+ },
+ {
+ "epoch": 40.23693379790941,
+ "grad_norm": 0.049527380615472794,
+ "learning_rate": 6.335028265666023e-05,
+ "loss": 1.7833847999572754,
+ "step": 2897
+ },
+ {
+ "epoch": 40.250871080139376,
+ "grad_norm": 0.04733026400208473,
+ "learning_rate": 6.319291022119653e-05,
+ "loss": 1.7732503414154053,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26480836236934,
+ "grad_norm": 0.04898470640182495,
+ "learning_rate": 6.303573099397378e-05,
+ "loss": 1.776735544204712,
+ "step": 2899
+ },
+ {
+ "epoch": 40.278745644599304,
+ "grad_norm": 0.04847757890820503,
+ "learning_rate": 6.287874510162821e-05,
+ "loss": 1.7859764099121094,
+ "step": 2900
+ },
+ {
+ "epoch": 40.292682926829265,
+ "grad_norm": 0.04857335984706879,
+ "learning_rate": 6.272195267064063e-05,
+ "loss": 1.7744362354278564,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30662020905923,
+ "grad_norm": 0.047604888677597046,
+ "learning_rate": 6.25653538273358e-05,
+ "loss": 1.7872700691223145,
+ "step": 2902
+ },
+ {
+ "epoch": 40.3205574912892,
+ "grad_norm": 0.048553720116615295,
+ "learning_rate": 6.240894869788267e-05,
+ "loss": 1.7771832942962646,
+ "step": 2903
+ },
+ {
+ "epoch": 40.33449477351916,
+ "grad_norm": 0.047940365970134735,
+ "learning_rate": 6.225273740829404e-05,
+ "loss": 1.7938380241394043,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34843205574913,
+ "grad_norm": 0.04888467490673065,
+ "learning_rate": 6.209672008442635e-05,
+ "loss": 1.7793422937393188,
+ "step": 2905
+ },
+ {
+ "epoch": 40.362369337979096,
+ "grad_norm": 0.04792848229408264,
+ "learning_rate": 6.19408968519801e-05,
+ "loss": 1.7843034267425537,
+ "step": 2906
+ },
+ {
+ "epoch": 40.37630662020906,
+ "grad_norm": 0.04912486672401428,
+ "learning_rate": 6.178526783649916e-05,
+ "loss": 1.7690515518188477,
+ "step": 2907
+ },
+ {
+ "epoch": 40.390243902439025,
+ "grad_norm": 0.04807637259364128,
+ "learning_rate": 6.162983316337109e-05,
+ "loss": 1.7809380292892456,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40418118466899,
+ "grad_norm": 0.04979358986020088,
+ "learning_rate": 6.147459295782676e-05,
+ "loss": 1.783470869064331,
+ "step": 2909
+ },
+ {
+ "epoch": 40.41811846689895,
+ "grad_norm": 0.049359280616045,
+ "learning_rate": 6.131954734494045e-05,
+ "loss": 1.7851728200912476,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43205574912892,
+ "grad_norm": 0.04928767681121826,
+ "learning_rate": 6.11646964496296e-05,
+ "loss": 1.7869231700897217,
+ "step": 2911
+ },
+ {
+ "epoch": 40.44599303135889,
+ "grad_norm": 0.04966576769948006,
+ "learning_rate": 6.101004039665471e-05,
+ "loss": 1.771344542503357,
+ "step": 2912
+ },
+ {
+ "epoch": 40.45993031358885,
+ "grad_norm": 0.04966041445732117,
+ "learning_rate": 6.085557931061937e-05,
+ "loss": 1.7944121360778809,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47386759581882,
+ "grad_norm": 0.04922773316502571,
+ "learning_rate": 6.070131331597015e-05,
+ "loss": 1.798032522201538,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48780487804878,
+ "grad_norm": 0.04820060357451439,
+ "learning_rate": 6.054724253699636e-05,
+ "loss": 1.7730165719985962,
+ "step": 2915
+ },
+ {
+ "epoch": 40.501742160278745,
+ "grad_norm": 0.04986457899212837,
+ "learning_rate": 6.03933670978301e-05,
+ "loss": 1.7808765172958374,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51567944250871,
+ "grad_norm": 0.04911157861351967,
+ "learning_rate": 6.0239687122445826e-05,
+ "loss": 1.7903361320495605,
+ "step": 2917
+ },
+ {
+ "epoch": 40.52961672473867,
+ "grad_norm": 0.048055119812488556,
+ "learning_rate": 6.0086202734661e-05,
+ "loss": 1.7886831760406494,
+ "step": 2918
+ },
+ {
+ "epoch": 40.54355400696864,
+ "grad_norm": 0.049517903476953506,
+ "learning_rate": 5.9932914058135004e-05,
+ "loss": 1.7857229709625244,
+ "step": 2919
+ },
+ {
+ "epoch": 40.55749128919861,
+ "grad_norm": 0.048604514449834824,
+ "learning_rate": 5.977982121636984e-05,
+ "loss": 1.7922042608261108,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57142857142857,
+ "grad_norm": 0.04886873811483383,
+ "learning_rate": 5.962692433270962e-05,
+ "loss": 1.7781345844268799,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58536585365854,
+ "grad_norm": 0.04913982376456261,
+ "learning_rate": 5.9474223530340634e-05,
+ "loss": 1.7909157276153564,
+ "step": 2922
+ },
+ {
+ "epoch": 40.599303135888505,
+ "grad_norm": 0.04856523871421814,
+ "learning_rate": 5.932171893229124e-05,
+ "loss": 1.7691776752471924,
+ "step": 2923
+ },
+ {
+ "epoch": 40.613240418118465,
+ "grad_norm": 0.05121750012040138,
+ "learning_rate": 5.916941066143137e-05,
+ "loss": 1.7849647998809814,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62717770034843,
+ "grad_norm": 0.04991038516163826,
+ "learning_rate": 5.9017298840473364e-05,
+ "loss": 1.7836909294128418,
+ "step": 2925
+ },
+ {
+ "epoch": 40.641114982578394,
+ "grad_norm": 0.04965447261929512,
+ "learning_rate": 5.8865383591970775e-05,
+ "loss": 1.7835614681243896,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65505226480836,
+ "grad_norm": 0.049823079258203506,
+ "learning_rate": 5.8713665038319e-05,
+ "loss": 1.7881591320037842,
+ "step": 2927
+ },
+ {
+ "epoch": 40.66898954703833,
+ "grad_norm": 0.049288421869277954,
+ "learning_rate": 5.856214330175498e-05,
+ "loss": 1.7872439622879028,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68292682926829,
+ "grad_norm": 0.05006309971213341,
+ "learning_rate": 5.841081850435704e-05,
+ "loss": 1.776764988899231,
+ "step": 2929
+ },
+ {
+ "epoch": 40.69686411149826,
+ "grad_norm": 0.05022395774722099,
+ "learning_rate": 5.825969076804488e-05,
+ "loss": 1.8023052215576172,
+ "step": 2930
+ },
+ {
+ "epoch": 40.710801393728225,
+ "grad_norm": 0.050583839416503906,
+ "learning_rate": 5.810876021457925e-05,
+ "loss": 1.783090353012085,
+ "step": 2931
+ },
+ {
+ "epoch": 40.724738675958186,
+ "grad_norm": 0.04937582463026047,
+ "learning_rate": 5.795802696556222e-05,
+ "loss": 1.7982478141784668,
+ "step": 2932
+ },
+ {
+ "epoch": 40.73867595818815,
+ "grad_norm": 0.04985835775732994,
+ "learning_rate": 5.78074911424369e-05,
+ "loss": 1.7760167121887207,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75261324041812,
+ "grad_norm": 0.05028945952653885,
+ "learning_rate": 5.765715286648734e-05,
+ "loss": 1.7793045043945312,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76655052264808,
+ "grad_norm": 0.05060959234833717,
+ "learning_rate": 5.7507012258838154e-05,
+ "loss": 1.7912629842758179,
+ "step": 2935
+ },
+ {
+ "epoch": 40.78048780487805,
+ "grad_norm": 0.05209024250507355,
+ "learning_rate": 5.7357069440455095e-05,
+ "loss": 1.7781007289886475,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79442508710802,
+ "grad_norm": 0.04997926205396652,
+ "learning_rate": 5.7207324532144454e-05,
+ "loss": 1.7866160869598389,
+ "step": 2937
+ },
+ {
+ "epoch": 40.80836236933798,
+ "grad_norm": 0.05141197517514229,
+ "learning_rate": 5.7057777654552785e-05,
+ "loss": 1.7921373844146729,
+ "step": 2938
+ },
+ {
+ "epoch": 40.822299651567945,
+ "grad_norm": 0.05031934380531311,
+ "learning_rate": 5.690842892816741e-05,
+ "loss": 1.786409854888916,
+ "step": 2939
+ },
+ {
+ "epoch": 40.836236933797906,
+ "grad_norm": 0.049375422298908234,
+ "learning_rate": 5.675927847331593e-05,
+ "loss": 1.7676091194152832,
+ "step": 2940
+ },
+ {
+ "epoch": 40.850174216027874,
+ "grad_norm": 0.05016583576798439,
+ "learning_rate": 5.661032641016619e-05,
+ "loss": 1.7917009592056274,
+ "step": 2941
+ },
+ {
+ "epoch": 40.86411149825784,
+ "grad_norm": 0.04995865747332573,
+ "learning_rate": 5.6461572858726086e-05,
+ "loss": 1.8048253059387207,
+ "step": 2942
+ },
+ {
+ "epoch": 40.8780487804878,
+ "grad_norm": 0.050674330443143845,
+ "learning_rate": 5.6313017938843595e-05,
+ "loss": 1.7829941511154175,
+ "step": 2943
+ },
+ {
+ "epoch": 40.89198606271777,
+ "grad_norm": 0.05021820589900017,
+ "learning_rate": 5.6164661770207026e-05,
+ "loss": 1.7910025119781494,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90592334494774,
+ "grad_norm": 0.05094406381249428,
+ "learning_rate": 5.6016504472344e-05,
+ "loss": 1.7892206907272339,
+ "step": 2945
+ },
+ {
+ "epoch": 40.9198606271777,
+ "grad_norm": 0.05145988613367081,
+ "learning_rate": 5.5868546164622284e-05,
+ "loss": 1.783717155456543,
+ "step": 2946
+ },
+ {
+ "epoch": 40.933797909407666,
+ "grad_norm": 0.05058317258954048,
+ "learning_rate": 5.572078696624917e-05,
+ "loss": 1.7758351564407349,
+ "step": 2947
+ },
+ {
+ "epoch": 40.94773519163763,
+ "grad_norm": 0.0512106828391552,
+ "learning_rate": 5.557322699627168e-05,
+ "loss": 1.8036705255508423,
+ "step": 2948
+ },
+ {
+ "epoch": 40.961672473867594,
+ "grad_norm": 0.0501241534948349,
+ "learning_rate": 5.542586637357607e-05,
+ "loss": 1.7745720148086548,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97560975609756,
+ "grad_norm": 0.05166011303663254,
+ "learning_rate": 5.527870521688817e-05,
+ "loss": 1.7932977676391602,
+ "step": 2950
+ },
+ {
+ "epoch": 40.98954703832753,
+ "grad_norm": 0.05050530284643173,
+ "learning_rate": 5.513174364477311e-05,
+ "loss": 1.7823102474212646,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.04222384840250015,
+ "learning_rate": 5.498498177563518e-05,
+ "loss": 1.3391203880310059,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 0.6519365310668945,
+ "eval_runtime": 40.5579,
+ "eval_samples_per_second": 60.21,
+ "eval_steps_per_second": 0.493,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01393728222997,
+ "grad_norm": 0.04689764603972435,
+ "learning_rate": 5.4838419727717716e-05,
+ "loss": 1.7729463577270508,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02787456445993,
+ "grad_norm": 0.044540271162986755,
+ "learning_rate": 5.4692057619103144e-05,
+ "loss": 1.7623910903930664,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041811846689896,
+ "grad_norm": 0.04553588107228279,
+ "learning_rate": 5.454589556771286e-05,
+ "loss": 1.7855658531188965,
+ "step": 2955
+ },
+ {
+ "epoch": 41.055749128919864,
+ "grad_norm": 0.04392184317111969,
+ "learning_rate": 5.439993369130681e-05,
+ "loss": 1.7711176872253418,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069686411149824,
+ "grad_norm": 0.04417131096124649,
+ "learning_rate": 5.4254172107483924e-05,
+ "loss": 1.7741881608963013,
+ "step": 2957
+ },
+ {
+ "epoch": 41.08362369337979,
+ "grad_norm": 0.04466690495610237,
+ "learning_rate": 5.410861093368169e-05,
+ "loss": 1.7754600048065186,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09756097560975,
+ "grad_norm": 0.045785821974277496,
+ "learning_rate": 5.3963250287176094e-05,
+ "loss": 1.7923266887664795,
+ "step": 2959
+ },
+ {
+ "epoch": 41.11149825783972,
+ "grad_norm": 0.044646237045526505,
+ "learning_rate": 5.381809028508168e-05,
+ "loss": 1.768218994140625,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12543554006969,
+ "grad_norm": 0.04554557055234909,
+ "learning_rate": 5.367313104435104e-05,
+ "loss": 1.7587878704071045,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13937282229965,
+ "grad_norm": 0.0454886369407177,
+ "learning_rate": 5.352837268177544e-05,
+ "loss": 1.7782719135284424,
+ "step": 2962
+ },
+ {
+ "epoch": 41.153310104529616,
+ "grad_norm": 0.045426227152347565,
+ "learning_rate": 5.3383815313983916e-05,
+ "loss": 1.775224208831787,
+ "step": 2963
+ },
+ {
+ "epoch": 41.167247386759584,
+ "grad_norm": 0.0466192252933979,
+ "learning_rate": 5.323945905744379e-05,
+ "loss": 1.7643415927886963,
+ "step": 2964
+ },
+ {
+ "epoch": 41.181184668989545,
+ "grad_norm": 0.045661959797143936,
+ "learning_rate": 5.3095304028460284e-05,
+ "loss": 1.7602108716964722,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19512195121951,
+ "grad_norm": 0.04575974866747856,
+ "learning_rate": 5.2951350343176495e-05,
+ "loss": 1.7654622793197632,
+ "step": 2966
+ },
+ {
+ "epoch": 41.20905923344948,
+ "grad_norm": 0.04690549150109291,
+ "learning_rate": 5.2807598117573384e-05,
+ "loss": 1.7760295867919922,
+ "step": 2967
+ },
+ {
+ "epoch": 41.22299651567944,
+ "grad_norm": 0.04673248901963234,
+ "learning_rate": 5.266404746746939e-05,
+ "loss": 1.7847099304199219,
+ "step": 2968
+ },
+ {
+ "epoch": 41.23693379790941,
+ "grad_norm": 0.04652085155248642,
+ "learning_rate": 5.252069850852072e-05,
+ "loss": 1.7894890308380127,
+ "step": 2969
+ },
+ {
+ "epoch": 41.250871080139376,
+ "grad_norm": 0.04779936745762825,
+ "learning_rate": 5.237755135622107e-05,
+ "loss": 1.7647104263305664,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26480836236934,
+ "grad_norm": 0.04592272266745567,
+ "learning_rate": 5.223460612590159e-05,
+ "loss": 1.7805603742599487,
+ "step": 2971
+ },
+ {
+ "epoch": 41.278745644599304,
+ "grad_norm": 0.0465652234852314,
+ "learning_rate": 5.209186293273044e-05,
+ "loss": 1.7796931266784668,
+ "step": 2972
+ },
+ {
+ "epoch": 41.292682926829265,
+ "grad_norm": 0.04658602178096771,
+ "learning_rate": 5.1949321891713446e-05,
+ "loss": 1.7612698078155518,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30662020905923,
+ "grad_norm": 0.04615809768438339,
+ "learning_rate": 5.180698311769338e-05,
+ "loss": 1.7831511497497559,
+ "step": 2974
+ },
+ {
+ "epoch": 41.3205574912892,
+ "grad_norm": 0.04704147204756737,
+ "learning_rate": 5.166484672534983e-05,
+ "loss": 1.7652500867843628,
+ "step": 2975
+ },
+ {
+ "epoch": 41.33449477351916,
+ "grad_norm": 0.046908169984817505,
+ "learning_rate": 5.152291282919961e-05,
+ "loss": 1.7832311391830444,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34843205574913,
+ "grad_norm": 0.046408504247665405,
+ "learning_rate": 5.138118154359632e-05,
+ "loss": 1.7829337120056152,
+ "step": 2977
+ },
+ {
+ "epoch": 41.362369337979096,
+ "grad_norm": 0.04615022614598274,
+ "learning_rate": 5.1239652982730335e-05,
+ "loss": 1.758888840675354,
+ "step": 2978
+ },
+ {
+ "epoch": 41.37630662020906,
+ "grad_norm": 0.046176742762327194,
+ "learning_rate": 5.1098327260628466e-05,
+ "loss": 1.7705339193344116,
+ "step": 2979
+ },
+ {
+ "epoch": 41.390243902439025,
+ "grad_norm": 0.048181675374507904,
+ "learning_rate": 5.09572044911545e-05,
+ "loss": 1.7633423805236816,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40418118466899,
+ "grad_norm": 0.047566913068294525,
+ "learning_rate": 5.0816284788008527e-05,
+ "loss": 1.7715733051300049,
+ "step": 2981
+ },
+ {
+ "epoch": 41.41811846689895,
+ "grad_norm": 0.046077970415353775,
+ "learning_rate": 5.067556826472684e-05,
+ "loss": 1.7751054763793945,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43205574912892,
+ "grad_norm": 0.045808132737874985,
+ "learning_rate": 5.053505503468228e-05,
+ "loss": 1.7841391563415527,
+ "step": 2983
+ },
+ {
+ "epoch": 41.44599303135889,
+ "grad_norm": 0.04701954126358032,
+ "learning_rate": 5.0394745211083785e-05,
+ "loss": 1.7742592096328735,
+ "step": 2984
+ },
+ {
+ "epoch": 41.45993031358885,
+ "grad_norm": 0.04649316146969795,
+ "learning_rate": 5.025463890697655e-05,
+ "loss": 1.7765004634857178,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47386759581882,
+ "grad_norm": 0.046931155025959015,
+ "learning_rate": 5.011473623524159e-05,
+ "loss": 1.780288815498352,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48780487804878,
+ "grad_norm": 0.04566241428256035,
+ "learning_rate": 4.9975037308595864e-05,
+ "loss": 1.781214952468872,
+ "step": 2987
+ },
+ {
+ "epoch": 41.501742160278745,
+ "grad_norm": 0.04714824631810188,
+ "learning_rate": 4.983554223959257e-05,
+ "loss": 1.77776038646698,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51567944250871,
+ "grad_norm": 0.046437617391347885,
+ "learning_rate": 4.9696251140620134e-05,
+ "loss": 1.764427900314331,
+ "step": 2989
+ },
+ {
+ "epoch": 41.52961672473867,
+ "grad_norm": 0.047795433551073074,
+ "learning_rate": 4.9557164123902924e-05,
+ "loss": 1.7826989889144897,
+ "step": 2990
+ },
+ {
+ "epoch": 41.54355400696864,
+ "grad_norm": 0.047585342079401016,
+ "learning_rate": 4.9418281301500844e-05,
+ "loss": 1.791701316833496,
+ "step": 2991
+ },
+ {
+ "epoch": 41.55749128919861,
+ "grad_norm": 0.046824004501104355,
+ "learning_rate": 4.9279602785309365e-05,
+ "loss": 1.781235933303833,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57142857142857,
+ "grad_norm": 0.04806610569357872,
+ "learning_rate": 4.914112868705911e-05,
+ "loss": 1.7798388004302979,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58536585365854,
+ "grad_norm": 0.04673585668206215,
+ "learning_rate": 4.900285911831624e-05,
+ "loss": 1.7777440547943115,
+ "step": 2994
+ },
+ {
+ "epoch": 41.599303135888505,
+ "grad_norm": 0.04765695706009865,
+ "learning_rate": 4.886479419048201e-05,
+ "loss": 1.771855115890503,
+ "step": 2995
+ },
+ {
+ "epoch": 41.613240418118465,
+ "grad_norm": 0.04678108170628548,
+ "learning_rate": 4.872693401479292e-05,
+ "loss": 1.7572466135025024,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62717770034843,
+ "grad_norm": 0.04738544300198555,
+ "learning_rate": 4.858927870232039e-05,
+ "loss": 1.7798383235931396,
+ "step": 2997
+ },
+ {
+ "epoch": 41.641114982578394,
+ "grad_norm": 0.048733294010162354,
+ "learning_rate": 4.845182836397068e-05,
+ "loss": 1.7774200439453125,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65505226480836,
+ "grad_norm": 0.046759042888879776,
+ "learning_rate": 4.831458311048532e-05,
+ "loss": 1.777611255645752,
+ "step": 2999
+ },
+ {
+ "epoch": 41.66898954703833,
+ "grad_norm": 0.04731730744242668,
+ "learning_rate": 4.817754305244008e-05,
+ "loss": 1.7866392135620117,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68292682926829,
+ "grad_norm": 0.04734078422188759,
+ "learning_rate": 4.804070830024577e-05,
+ "loss": 1.7836531400680542,
+ "step": 3001
+ },
+ {
+ "epoch": 41.69686411149826,
+ "grad_norm": 0.04705912247300148,
+ "learning_rate": 4.7904078964147654e-05,
+ "loss": 1.7780985832214355,
+ "step": 3002
+ },
+ {
+ "epoch": 41.710801393728225,
+ "grad_norm": 0.048269402235746384,
+ "learning_rate": 4.776765515422557e-05,
+ "loss": 1.7863717079162598,
+ "step": 3003
+ },
+ {
+ "epoch": 41.724738675958186,
+ "grad_norm": 0.0475349985063076,
+ "learning_rate": 4.763143698039371e-05,
+ "loss": 1.7757699489593506,
+ "step": 3004
+ },
+ {
+ "epoch": 41.73867595818815,
+ "grad_norm": 0.0482979416847229,
+ "learning_rate": 4.749542455240043e-05,
+ "loss": 1.7803300619125366,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75261324041812,
+ "grad_norm": 0.0486641526222229,
+ "learning_rate": 4.735961797982876e-05,
+ "loss": 1.7695112228393555,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76655052264808,
+ "grad_norm": 0.048009518533945084,
+ "learning_rate": 4.7224017372095414e-05,
+ "loss": 1.7627280950546265,
+ "step": 3007
+ },
+ {
+ "epoch": 41.78048780487805,
+ "grad_norm": 0.04899517074227333,
+ "learning_rate": 4.708862283845143e-05,
+ "loss": 1.7776708602905273,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79442508710802,
+ "grad_norm": 0.047490231692790985,
+ "learning_rate": 4.695343448798163e-05,
+ "loss": 1.7935547828674316,
+ "step": 3009
+ },
+ {
+ "epoch": 41.80836236933798,
+ "grad_norm": 0.04828681796789169,
+ "learning_rate": 4.681845242960492e-05,
+ "loss": 1.7704858779907227,
+ "step": 3010
+ },
+ {
+ "epoch": 41.822299651567945,
+ "grad_norm": 0.04789802059531212,
+ "learning_rate": 4.668367677207398e-05,
+ "loss": 1.7866098880767822,
+ "step": 3011
+ },
+ {
+ "epoch": 41.836236933797906,
+ "grad_norm": 0.04726841300725937,
+ "learning_rate": 4.654910762397499e-05,
+ "loss": 1.775240421295166,
+ "step": 3012
+ },
+ {
+ "epoch": 41.850174216027874,
+ "grad_norm": 0.048522237688302994,
+ "learning_rate": 4.6414745093727913e-05,
+ "loss": 1.7788512706756592,
+ "step": 3013
+ },
+ {
+ "epoch": 41.86411149825784,
+ "grad_norm": 0.04785652831196785,
+ "learning_rate": 4.6280589289586255e-05,
+ "loss": 1.7695248126983643,
+ "step": 3014
+ },
+ {
+ "epoch": 41.8780487804878,
+ "grad_norm": 0.04859865456819534,
+ "learning_rate": 4.614664031963692e-05,
+ "loss": 1.768019437789917,
+ "step": 3015
+ },
+ {
+ "epoch": 41.89198606271777,
+ "grad_norm": 0.04948209971189499,
+ "learning_rate": 4.601289829180004e-05,
+ "loss": 1.7800569534301758,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90592334494774,
+ "grad_norm": 0.04760837182402611,
+ "learning_rate": 4.587936331382934e-05,
+ "loss": 1.7688634395599365,
+ "step": 3017
+ },
+ {
+ "epoch": 41.9198606271777,
+ "grad_norm": 0.04870005324482918,
+ "learning_rate": 4.574603549331151e-05,
+ "loss": 1.7903432846069336,
+ "step": 3018
+ },
+ {
+ "epoch": 41.933797909407666,
+ "grad_norm": 0.04837913066148758,
+ "learning_rate": 4.561291493766625e-05,
+ "loss": 1.7871191501617432,
+ "step": 3019
+ },
+ {
+ "epoch": 41.94773519163763,
+ "grad_norm": 0.048451390117406845,
+ "learning_rate": 4.5480001754146455e-05,
+ "loss": 1.7867250442504883,
+ "step": 3020
+ },
+ {
+ "epoch": 41.961672473867594,
+ "grad_norm": 0.04890143498778343,
+ "learning_rate": 4.5347296049837875e-05,
+ "loss": 1.7735885381698608,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97560975609756,
+ "grad_norm": 0.04721386358141899,
+ "learning_rate": 4.52147979316592e-05,
+ "loss": 1.7692224979400635,
+ "step": 3022
+ },
+ {
+ "epoch": 41.98954703832753,
+ "grad_norm": 0.04814184084534645,
+ "learning_rate": 4.5082507506361574e-05,
+ "loss": 1.776921033859253,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.039818305522203445,
+ "learning_rate": 4.495042488052904e-05,
+ "loss": 1.3309056758880615,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 0.6534227132797241,
+ "eval_runtime": 40.6489,
+ "eval_samples_per_second": 60.075,
+ "eval_steps_per_second": 0.492,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01393728222997,
+ "grad_norm": 0.04341503232717514,
+ "learning_rate": 4.481855016057839e-05,
+ "loss": 1.7888715267181396,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02787456445993,
+ "grad_norm": 0.04266861081123352,
+ "learning_rate": 4.468688345275848e-05,
+ "loss": 1.772957682609558,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041811846689896,
+ "grad_norm": 0.043375164270401,
+ "learning_rate": 4.455542486315086e-05,
+ "loss": 1.7768115997314453,
+ "step": 3027
+ },
+ {
+ "epoch": 42.055749128919864,
+ "grad_norm": 0.044685594737529755,
+ "learning_rate": 4.442417449766938e-05,
+ "loss": 1.7511389255523682,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069686411149824,
+ "grad_norm": 0.0430758111178875,
+ "learning_rate": 4.429313246206014e-05,
+ "loss": 1.7582142353057861,
+ "step": 3029
+ },
+ {
+ "epoch": 42.08362369337979,
+ "grad_norm": 0.04426584020256996,
+ "learning_rate": 4.416229886190119e-05,
+ "loss": 1.7670388221740723,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09756097560975,
+ "grad_norm": 0.0441875085234642,
+ "learning_rate": 4.403167380260281e-05,
+ "loss": 1.7622346878051758,
+ "step": 3031
+ },
+ {
+ "epoch": 42.11149825783972,
+ "grad_norm": 0.04547901079058647,
+ "learning_rate": 4.3901257389407475e-05,
+ "loss": 1.7697657346725464,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12543554006969,
+ "grad_norm": 0.04314858466386795,
+ "learning_rate": 4.3771049727389075e-05,
+ "loss": 1.7717976570129395,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13937282229965,
+ "grad_norm": 0.04366912320256233,
+ "learning_rate": 4.364105092145377e-05,
+ "loss": 1.7681739330291748,
+ "step": 3034
+ },
+ {
+ "epoch": 42.153310104529616,
+ "grad_norm": 0.04399389028549194,
+ "learning_rate": 4.3511261076339e-05,
+ "loss": 1.763157844543457,
+ "step": 3035
+ },
+ {
+ "epoch": 42.167247386759584,
+ "grad_norm": 0.043772801756858826,
+ "learning_rate": 4.3381680296614394e-05,
+ "loss": 1.768390417098999,
+ "step": 3036
+ },
+ {
+ "epoch": 42.181184668989545,
+ "grad_norm": 0.04431698098778725,
+ "learning_rate": 4.3252308686680626e-05,
+ "loss": 1.7630479335784912,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19512195121951,
+ "grad_norm": 0.04417581111192703,
+ "learning_rate": 4.3123146350770146e-05,
+ "loss": 1.771895408630371,
+ "step": 3038
+ },
+ {
+ "epoch": 42.20905923344948,
+ "grad_norm": 0.045676663517951965,
+ "learning_rate": 4.2994193392946726e-05,
+ "loss": 1.7609918117523193,
+ "step": 3039
+ },
+ {
+ "epoch": 42.22299651567944,
+ "grad_norm": 0.04374372959136963,
+ "learning_rate": 4.286544991710541e-05,
+ "loss": 1.7676066160202026,
+ "step": 3040
+ },
+ {
+ "epoch": 42.23693379790941,
+ "grad_norm": 0.043877895921468735,
+ "learning_rate": 4.2736916026972576e-05,
+ "loss": 1.7593367099761963,
+ "step": 3041
+ },
+ {
+ "epoch": 42.250871080139376,
+ "grad_norm": 0.0448957122862339,
+ "learning_rate": 4.260859182610542e-05,
+ "loss": 1.7552883625030518,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26480836236934,
+ "grad_norm": 0.04429576173424721,
+ "learning_rate": 4.2480477417892776e-05,
+ "loss": 1.7635066509246826,
+ "step": 3043
+ },
+ {
+ "epoch": 42.278745644599304,
+ "grad_norm": 0.04319818317890167,
+ "learning_rate": 4.235257290555381e-05,
+ "loss": 1.7759575843811035,
+ "step": 3044
+ },
+ {
+ "epoch": 42.292682926829265,
+ "grad_norm": 0.04452989250421524,
+ "learning_rate": 4.222487839213903e-05,
+ "loss": 1.7780866622924805,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30662020905923,
+ "grad_norm": 0.044572584331035614,
+ "learning_rate": 4.209739398052956e-05,
+ "loss": 1.771972894668579,
+ "step": 3046
+ },
+ {
+ "epoch": 42.3205574912892,
+ "grad_norm": 0.044774096459150314,
+ "learning_rate": 4.1970119773437316e-05,
+ "loss": 1.7678005695343018,
+ "step": 3047
+ },
+ {
+ "epoch": 42.33449477351916,
+ "grad_norm": 0.0447823666036129,
+ "learning_rate": 4.184305587340483e-05,
+ "loss": 1.7687022686004639,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34843205574913,
+ "grad_norm": 0.04409811645746231,
+ "learning_rate": 4.171620238280511e-05,
+ "loss": 1.759227991104126,
+ "step": 3049
+ },
+ {
+ "epoch": 42.362369337979096,
+ "grad_norm": 0.04530395194888115,
+ "learning_rate": 4.158955940384179e-05,
+ "loss": 1.76583731174469,
+ "step": 3050
+ },
+ {
+ "epoch": 42.37630662020906,
+ "grad_norm": 0.043990686535835266,
+ "learning_rate": 4.146312703854881e-05,
+ "loss": 1.776308536529541,
+ "step": 3051
+ },
+ {
+ "epoch": 42.390243902439025,
+ "grad_norm": 0.045310501009225845,
+ "learning_rate": 4.133690538879046e-05,
+ "loss": 1.7690048217773438,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40418118466899,
+ "grad_norm": 0.045838840305805206,
+ "learning_rate": 4.1210894556261226e-05,
+ "loss": 1.7719197273254395,
+ "step": 3053
+ },
+ {
+ "epoch": 42.41811846689895,
+ "grad_norm": 0.044539984315633774,
+ "learning_rate": 4.108509464248578e-05,
+ "loss": 1.768972635269165,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43205574912892,
+ "grad_norm": 0.04546492174267769,
+ "learning_rate": 4.095950574881891e-05,
+ "loss": 1.7548890113830566,
+ "step": 3055
+ },
+ {
+ "epoch": 42.44599303135889,
+ "grad_norm": 0.0445762574672699,
+ "learning_rate": 4.083412797644514e-05,
+ "loss": 1.7676918506622314,
+ "step": 3056
+ },
+ {
+ "epoch": 42.45993031358885,
+ "grad_norm": 0.045404110103845596,
+ "learning_rate": 4.0708961426379214e-05,
+ "loss": 1.7771313190460205,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47386759581882,
+ "grad_norm": 0.04481915012001991,
+ "learning_rate": 4.0584006199465484e-05,
+ "loss": 1.7662057876586914,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48780487804878,
+ "grad_norm": 0.044174905866384506,
+ "learning_rate": 4.0459262396378165e-05,
+ "loss": 1.7512820959091187,
+ "step": 3059
+ },
+ {
+ "epoch": 42.501742160278745,
+ "grad_norm": 0.04443437233567238,
+ "learning_rate": 4.033473011762116e-05,
+ "loss": 1.76678466796875,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51567944250871,
+ "grad_norm": 0.04414868354797363,
+ "learning_rate": 4.0210409463527656e-05,
+ "loss": 1.7657546997070312,
+ "step": 3061
+ },
+ {
+ "epoch": 42.52961672473867,
+ "grad_norm": 0.045545708388090134,
+ "learning_rate": 4.008630053426082e-05,
+ "loss": 1.7497735023498535,
+ "step": 3062
+ },
+ {
+ "epoch": 42.54355400696864,
+ "grad_norm": 0.045831989496946335,
+ "learning_rate": 3.996240342981279e-05,
+ "loss": 1.7767338752746582,
+ "step": 3063
+ },
+ {
+ "epoch": 42.55749128919861,
+ "grad_norm": 0.0455077588558197,
+ "learning_rate": 3.983871825000529e-05,
+ "loss": 1.7681231498718262,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57142857142857,
+ "grad_norm": 0.045490484684705734,
+ "learning_rate": 3.971524509448925e-05,
+ "loss": 1.7701823711395264,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58536585365854,
+ "grad_norm": 0.04617610201239586,
+ "learning_rate": 3.959198406274472e-05,
+ "loss": 1.7651724815368652,
+ "step": 3066
+ },
+ {
+ "epoch": 42.599303135888505,
+ "grad_norm": 0.045694541186094284,
+ "learning_rate": 3.946893525408097e-05,
+ "loss": 1.7682925462722778,
+ "step": 3067
+ },
+ {
+ "epoch": 42.613240418118465,
+ "grad_norm": 0.045455750077962875,
+ "learning_rate": 3.934609876763604e-05,
+ "loss": 1.771888017654419,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62717770034843,
+ "grad_norm": 0.04597267881035805,
+ "learning_rate": 3.922347470237728e-05,
+ "loss": 1.776789665222168,
+ "step": 3069
+ },
+ {
+ "epoch": 42.641114982578394,
+ "grad_norm": 0.046098366379737854,
+ "learning_rate": 3.910106315710053e-05,
+ "loss": 1.7781219482421875,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65505226480836,
+ "grad_norm": 0.04562405124306679,
+ "learning_rate": 3.897886423043061e-05,
+ "loss": 1.7662599086761475,
+ "step": 3071
+ },
+ {
+ "epoch": 42.66898954703833,
+ "grad_norm": 0.0461348257958889,
+ "learning_rate": 3.8856878020820975e-05,
+ "loss": 1.775161623954773,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68292682926829,
+ "grad_norm": 0.04594029486179352,
+ "learning_rate": 3.873510462655375e-05,
+ "loss": 1.7788872718811035,
+ "step": 3073
+ },
+ {
+ "epoch": 42.69686411149826,
+ "grad_norm": 0.0456400141119957,
+ "learning_rate": 3.861354414573954e-05,
+ "loss": 1.7744836807250977,
+ "step": 3074
+ },
+ {
+ "epoch": 42.710801393728225,
+ "grad_norm": 0.045577649027109146,
+ "learning_rate": 3.849219667631735e-05,
+ "loss": 1.7715539932250977,
+ "step": 3075
+ },
+ {
+ "epoch": 42.724738675958186,
+ "grad_norm": 0.046013571321964264,
+ "learning_rate": 3.8371062316054764e-05,
+ "loss": 1.7707717418670654,
+ "step": 3076
+ },
+ {
+ "epoch": 42.73867595818815,
+ "grad_norm": 0.046062543988227844,
+ "learning_rate": 3.8250141162547466e-05,
+ "loss": 1.7822339534759521,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75261324041812,
+ "grad_norm": 0.04651106148958206,
+ "learning_rate": 3.812943331321956e-05,
+ "loss": 1.7592147588729858,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76655052264808,
+ "grad_norm": 0.046771664172410965,
+ "learning_rate": 3.8008938865322955e-05,
+ "loss": 1.7719390392303467,
+ "step": 3079
+ },
+ {
+ "epoch": 42.78048780487805,
+ "grad_norm": 0.04678458347916603,
+ "learning_rate": 3.788865791593813e-05,
+ "loss": 1.7596237659454346,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79442508710802,
+ "grad_norm": 0.046154703944921494,
+ "learning_rate": 3.7768590561973064e-05,
+ "loss": 1.7708245515823364,
+ "step": 3081
+ },
+ {
+ "epoch": 42.80836236933798,
+ "grad_norm": 0.04633764550089836,
+ "learning_rate": 3.764873690016394e-05,
+ "loss": 1.7710938453674316,
+ "step": 3082
+ },
+ {
+ "epoch": 42.822299651567945,
+ "grad_norm": 0.046283479779958725,
+ "learning_rate": 3.7529097027074654e-05,
+ "loss": 1.7665671110153198,
+ "step": 3083
+ },
+ {
+ "epoch": 42.836236933797906,
+ "grad_norm": 0.04648987576365471,
+ "learning_rate": 3.7409671039096936e-05,
+ "loss": 1.7542476654052734,
+ "step": 3084
+ },
+ {
+ "epoch": 42.850174216027874,
+ "grad_norm": 0.04672298952937126,
+ "learning_rate": 3.7290459032450184e-05,
+ "loss": 1.7689290046691895,
+ "step": 3085
+ },
+ {
+ "epoch": 42.86411149825784,
+ "grad_norm": 0.046370647847652435,
+ "learning_rate": 3.717146110318123e-05,
+ "loss": 1.7649741172790527,
+ "step": 3086
+ },
+ {
+ "epoch": 42.8780487804878,
+ "grad_norm": 0.047446198761463165,
+ "learning_rate": 3.7052677347164614e-05,
+ "loss": 1.7729518413543701,
+ "step": 3087
+ },
+ {
+ "epoch": 42.89198606271777,
+ "grad_norm": 0.04655609652400017,
+ "learning_rate": 3.693410786010231e-05,
+ "loss": 1.775625228881836,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90592334494774,
+ "grad_norm": 0.046239983290433884,
+ "learning_rate": 3.6815752737523537e-05,
+ "loss": 1.7640939950942993,
+ "step": 3089
+ },
+ {
+ "epoch": 42.9198606271777,
+ "grad_norm": 0.046714168041944504,
+ "learning_rate": 3.6697612074784944e-05,
+ "loss": 1.7750494480133057,
+ "step": 3090
+ },
+ {
+ "epoch": 42.933797909407666,
+ "grad_norm": 0.0479653924703598,
+ "learning_rate": 3.657968596707029e-05,
+ "loss": 1.7775356769561768,
+ "step": 3091
+ },
+ {
+ "epoch": 42.94773519163763,
+ "grad_norm": 0.0469873771071434,
+ "learning_rate": 3.64619745093906e-05,
+ "loss": 1.7633222341537476,
+ "step": 3092
+ },
+ {
+ "epoch": 42.961672473867594,
+ "grad_norm": 0.04614265635609627,
+ "learning_rate": 3.6344477796583714e-05,
+ "loss": 1.7667288780212402,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97560975609756,
+ "grad_norm": 0.04644224792718887,
+ "learning_rate": 3.622719592331472e-05,
+ "loss": 1.7725608348846436,
+ "step": 3094
+ },
+ {
+ "epoch": 42.98954703832753,
+ "grad_norm": 0.04625892639160156,
+ "learning_rate": 3.611012898407544e-05,
+ "loss": 1.7723143100738525,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.03922572359442711,
+ "learning_rate": 3.599327707318464e-05,
+ "loss": 1.3379738330841064,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 0.654710054397583,
+ "eval_runtime": 40.5041,
+ "eval_samples_per_second": 60.29,
+ "eval_steps_per_second": 0.494,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01393728222997,
+ "grad_norm": 0.042021218687295914,
+ "learning_rate": 3.587664028478779e-05,
+ "loss": 1.7534828186035156,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02787456445993,
+ "grad_norm": 0.04172177240252495,
+ "learning_rate": 3.5760218712857046e-05,
+ "loss": 1.7588303089141846,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041811846689896,
+ "grad_norm": 0.042629424482584,
+ "learning_rate": 3.564401245119119e-05,
+ "loss": 1.764153242111206,
+ "step": 3099
+ },
+ {
+ "epoch": 43.055749128919864,
+ "grad_norm": 0.04150637611746788,
+ "learning_rate": 3.552802159341543e-05,
+ "loss": 1.7654709815979004,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069686411149824,
+ "grad_norm": 0.042983464896678925,
+ "learning_rate": 3.541224623298155e-05,
+ "loss": 1.7520860433578491,
+ "step": 3101
+ },
+ {
+ "epoch": 43.08362369337979,
+ "grad_norm": 0.04234235733747482,
+ "learning_rate": 3.529668646316765e-05,
+ "loss": 1.7629724740982056,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09756097560975,
+ "grad_norm": 0.04364536330103874,
+ "learning_rate": 3.5181342377078166e-05,
+ "loss": 1.7700273990631104,
+ "step": 3103
+ },
+ {
+ "epoch": 43.11149825783972,
+ "grad_norm": 0.043138496577739716,
+ "learning_rate": 3.5066214067643785e-05,
+ "loss": 1.7651735544204712,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12543554006969,
+ "grad_norm": 0.041396476328372955,
+ "learning_rate": 3.495130162762119e-05,
+ "loss": 1.7515389919281006,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13937282229965,
+ "grad_norm": 0.04298311099410057,
+ "learning_rate": 3.483660514959343e-05,
+ "loss": 1.7651766538619995,
+ "step": 3106
+ },
+ {
+ "epoch": 43.153310104529616,
+ "grad_norm": 0.04275120422244072,
+ "learning_rate": 3.4722124725969235e-05,
+ "loss": 1.7559534311294556,
+ "step": 3107
+ },
+ {
+ "epoch": 43.167247386759584,
+ "grad_norm": 0.041940510272979736,
+ "learning_rate": 3.460786044898346e-05,
+ "loss": 1.7608134746551514,
+ "step": 3108
+ },
+ {
+ "epoch": 43.181184668989545,
+ "grad_norm": 0.043498437851667404,
+ "learning_rate": 3.4493812410696784e-05,
+ "loss": 1.7735629081726074,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19512195121951,
+ "grad_norm": 0.04355790466070175,
+ "learning_rate": 3.437998070299564e-05,
+ "loss": 1.7664821147918701,
+ "step": 3110
+ },
+ {
+ "epoch": 43.20905923344948,
+ "grad_norm": 0.042053356766700745,
+ "learning_rate": 3.426636541759225e-05,
+ "loss": 1.7596155405044556,
+ "step": 3111
+ },
+ {
+ "epoch": 43.22299651567944,
+ "grad_norm": 0.04320033639669418,
+ "learning_rate": 3.415296664602425e-05,
+ "loss": 1.7779096364974976,
+ "step": 3112
+ },
+ {
+ "epoch": 43.23693379790941,
+ "grad_norm": 0.042722251266241074,
+ "learning_rate": 3.403978447965507e-05,
+ "loss": 1.771970510482788,
+ "step": 3113
+ },
+ {
+ "epoch": 43.250871080139376,
+ "grad_norm": 0.043361593037843704,
+ "learning_rate": 3.392681900967355e-05,
+ "loss": 1.7600276470184326,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26480836236934,
+ "grad_norm": 0.04405055567622185,
+ "learning_rate": 3.381407032709395e-05,
+ "loss": 1.7721140384674072,
+ "step": 3115
+ },
+ {
+ "epoch": 43.278745644599304,
+ "grad_norm": 0.04226464778184891,
+ "learning_rate": 3.3701538522755714e-05,
+ "loss": 1.755659818649292,
+ "step": 3116
+ },
+ {
+ "epoch": 43.292682926829265,
+ "grad_norm": 0.04350192844867706,
+ "learning_rate": 3.35892236873238e-05,
+ "loss": 1.7596216201782227,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30662020905923,
+ "grad_norm": 0.04462328553199768,
+ "learning_rate": 3.3477125911288284e-05,
+ "loss": 1.7591629028320312,
+ "step": 3118
+ },
+ {
+ "epoch": 43.3205574912892,
+ "grad_norm": 0.04288632422685623,
+ "learning_rate": 3.3365245284964225e-05,
+ "loss": 1.7450718879699707,
+ "step": 3119
+ },
+ {
+ "epoch": 43.33449477351916,
+ "grad_norm": 0.0438922680914402,
+ "learning_rate": 3.3253581898491836e-05,
+ "loss": 1.765263557434082,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34843205574913,
+ "grad_norm": 0.04307923838496208,
+ "learning_rate": 3.3142135841836276e-05,
+ "loss": 1.7653264999389648,
+ "step": 3121
+ },
+ {
+ "epoch": 43.362369337979096,
+ "grad_norm": 0.04415101930499077,
+ "learning_rate": 3.303090720478773e-05,
+ "loss": 1.7691493034362793,
+ "step": 3122
+ },
+ {
+ "epoch": 43.37630662020906,
+ "grad_norm": 0.04457402974367142,
+ "learning_rate": 3.291989607696092e-05,
+ "loss": 1.7685725688934326,
+ "step": 3123
+ },
+ {
+ "epoch": 43.390243902439025,
+ "grad_norm": 0.04351523891091347,
+ "learning_rate": 3.280910254779565e-05,
+ "loss": 1.7489142417907715,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40418118466899,
+ "grad_norm": 0.043415915220975876,
+ "learning_rate": 3.269852670655629e-05,
+ "loss": 1.7557696104049683,
+ "step": 3125
+ },
+ {
+ "epoch": 43.41811846689895,
+ "grad_norm": 0.044516436755657196,
+ "learning_rate": 3.258816864233169e-05,
+ "loss": 1.7655267715454102,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43205574912892,
+ "grad_norm": 0.04449203237891197,
+ "learning_rate": 3.247802844403536e-05,
+ "loss": 1.7631142139434814,
+ "step": 3127
+ },
+ {
+ "epoch": 43.44599303135889,
+ "grad_norm": 0.04321179911494255,
+ "learning_rate": 3.236810620040532e-05,
+ "loss": 1.7660303115844727,
+ "step": 3128
+ },
+ {
+ "epoch": 43.45993031358885,
+ "grad_norm": 0.04363849759101868,
+ "learning_rate": 3.225840200000396e-05,
+ "loss": 1.75204598903656,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47386759581882,
+ "grad_norm": 0.04379528760910034,
+ "learning_rate": 3.214891593121788e-05,
+ "loss": 1.7515747547149658,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48780487804878,
+ "grad_norm": 0.04358968138694763,
+ "learning_rate": 3.203964808225801e-05,
+ "loss": 1.7563517093658447,
+ "step": 3131
+ },
+ {
+ "epoch": 43.501742160278745,
+ "grad_norm": 0.043669551610946655,
+ "learning_rate": 3.193059854115968e-05,
+ "loss": 1.762972354888916,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51567944250871,
+ "grad_norm": 0.04360586032271385,
+ "learning_rate": 3.182176739578195e-05,
+ "loss": 1.7609463930130005,
+ "step": 3133
+ },
+ {
+ "epoch": 43.52961672473867,
+ "grad_norm": 0.04505139961838722,
+ "learning_rate": 3.171315473380815e-05,
+ "loss": 1.7723445892333984,
+ "step": 3134
+ },
+ {
+ "epoch": 43.54355400696864,
+ "grad_norm": 0.043562229722738266,
+ "learning_rate": 3.160476064274555e-05,
+ "loss": 1.7522692680358887,
+ "step": 3135
+ },
+ {
+ "epoch": 43.55749128919861,
+ "grad_norm": 0.04364638775587082,
+ "learning_rate": 3.149658520992537e-05,
+ "loss": 1.7644182443618774,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57142857142857,
+ "grad_norm": 0.04494662210345268,
+ "learning_rate": 3.138862852250249e-05,
+ "loss": 1.7732231616973877,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58536585365854,
+ "grad_norm": 0.04464591667056084,
+ "learning_rate": 3.1280890667455694e-05,
+ "loss": 1.754976749420166,
+ "step": 3138
+ },
+ {
+ "epoch": 43.599303135888505,
+ "grad_norm": 0.04465331509709358,
+ "learning_rate": 3.1173371731587496e-05,
+ "loss": 1.7709685564041138,
+ "step": 3139
+ },
+ {
+ "epoch": 43.613240418118465,
+ "grad_norm": 0.04342262074351311,
+ "learning_rate": 3.1066071801523905e-05,
+ "loss": 1.7601979970932007,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62717770034843,
+ "grad_norm": 0.04447892680764198,
+ "learning_rate": 3.095899096371456e-05,
+ "loss": 1.756757378578186,
+ "step": 3141
+ },
+ {
+ "epoch": 43.641114982578394,
+ "grad_norm": 0.04505394026637077,
+ "learning_rate": 3.0852129304432506e-05,
+ "loss": 1.7673622369766235,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65505226480836,
+ "grad_norm": 0.044277530163526535,
+ "learning_rate": 3.074548690977434e-05,
+ "loss": 1.762864112854004,
+ "step": 3143
+ },
+ {
+ "epoch": 43.66898954703833,
+ "grad_norm": 0.04379425197839737,
+ "learning_rate": 3.0639063865659834e-05,
+ "loss": 1.7678565979003906,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68292682926829,
+ "grad_norm": 0.04339452087879181,
+ "learning_rate": 3.0532860257832144e-05,
+ "loss": 1.7612366676330566,
+ "step": 3145
+ },
+ {
+ "epoch": 43.69686411149826,
+ "grad_norm": 0.04434647411108017,
+ "learning_rate": 3.04268761718576e-05,
+ "loss": 1.7668228149414062,
+ "step": 3146
+ },
+ {
+ "epoch": 43.710801393728225,
+ "grad_norm": 0.04466056823730469,
+ "learning_rate": 3.0321111693125648e-05,
+ "loss": 1.7554584741592407,
+ "step": 3147
+ },
+ {
+ "epoch": 43.724738675958186,
+ "grad_norm": 0.044375862926244736,
+ "learning_rate": 3.0215566906848906e-05,
+ "loss": 1.7701786756515503,
+ "step": 3148
+ },
+ {
+ "epoch": 43.73867595818815,
+ "grad_norm": 0.044433314353227615,
+ "learning_rate": 3.0110241898062726e-05,
+ "loss": 1.7586554288864136,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75261324041812,
+ "grad_norm": 0.0449117012321949,
+ "learning_rate": 3.0005136751625748e-05,
+ "loss": 1.7642850875854492,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76655052264808,
+ "grad_norm": 0.04465561732649803,
+ "learning_rate": 2.9900251552219178e-05,
+ "loss": 1.7654483318328857,
+ "step": 3151
+ },
+ {
+ "epoch": 43.78048780487805,
+ "grad_norm": 0.044917911291122437,
+ "learning_rate": 2.9795586384347202e-05,
+ "loss": 1.7696160078048706,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79442508710802,
+ "grad_norm": 0.044988881796598434,
+ "learning_rate": 2.9691141332336494e-05,
+ "loss": 1.7712256908416748,
+ "step": 3153
+ },
+ {
+ "epoch": 43.80836236933798,
+ "grad_norm": 0.04482322186231613,
+ "learning_rate": 2.9586916480336735e-05,
+ "loss": 1.762773871421814,
+ "step": 3154
+ },
+ {
+ "epoch": 43.822299651567945,
+ "grad_norm": 0.04423758387565613,
+ "learning_rate": 2.9482911912320006e-05,
+ "loss": 1.7609460353851318,
+ "step": 3155
+ },
+ {
+ "epoch": 43.836236933797906,
+ "grad_norm": 0.044506512582302094,
+ "learning_rate": 2.9379127712080756e-05,
+ "loss": 1.765155553817749,
+ "step": 3156
+ },
+ {
+ "epoch": 43.850174216027874,
+ "grad_norm": 0.04564986377954483,
+ "learning_rate": 2.9275563963236107e-05,
+ "loss": 1.7608258724212646,
+ "step": 3157
+ },
+ {
+ "epoch": 43.86411149825784,
+ "grad_norm": 0.043920136988162994,
+ "learning_rate": 2.9172220749225548e-05,
+ "loss": 1.76609468460083,
+ "step": 3158
+ },
+ {
+ "epoch": 43.8780487804878,
+ "grad_norm": 0.04496336728334427,
+ "learning_rate": 2.906909815331085e-05,
+ "loss": 1.761902093887329,
+ "step": 3159
+ },
+ {
+ "epoch": 43.89198606271777,
+ "grad_norm": 0.0450543649494648,
+ "learning_rate": 2.896619625857588e-05,
+ "loss": 1.7570101022720337,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90592334494774,
+ "grad_norm": 0.04378526285290718,
+ "learning_rate": 2.8863515147927003e-05,
+ "loss": 1.7584224939346313,
+ "step": 3161
+ },
+ {
+ "epoch": 43.9198606271777,
+ "grad_norm": 0.04425406455993652,
+ "learning_rate": 2.8761054904092514e-05,
+ "loss": 1.7499665021896362,
+ "step": 3162
+ },
+ {
+ "epoch": 43.933797909407666,
+ "grad_norm": 0.044044118374586105,
+ "learning_rate": 2.8658815609622665e-05,
+ "loss": 1.739232063293457,
+ "step": 3163
+ },
+ {
+ "epoch": 43.94773519163763,
+ "grad_norm": 0.043724123388528824,
+ "learning_rate": 2.855679734688993e-05,
+ "loss": 1.7675073146820068,
+ "step": 3164
+ },
+ {
+ "epoch": 43.961672473867594,
+ "grad_norm": 0.04547949880361557,
+ "learning_rate": 2.8455000198088502e-05,
+ "loss": 1.7621852159500122,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97560975609756,
+ "grad_norm": 0.04585639387369156,
+ "learning_rate": 2.8353424245234582e-05,
+ "loss": 1.759611964225769,
+ "step": 3166
+ },
+ {
+ "epoch": 43.98954703832753,
+ "grad_norm": 0.04513043910264969,
+ "learning_rate": 2.825206957016599e-05,
+ "loss": 1.7627670764923096,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.037081386893987656,
+ "learning_rate": 2.8150936254542376e-05,
+ "loss": 1.3273016214370728,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 0.655878484249115,
+ "eval_runtime": 62.6925,
+ "eval_samples_per_second": 38.952,
+ "eval_steps_per_second": 0.319,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01393728222997,
+ "grad_norm": 0.04096479341387749,
+ "learning_rate": 2.8050024379845118e-05,
+ "loss": 1.7576605081558228,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02787456445993,
+ "grad_norm": 0.040468111634254456,
+ "learning_rate": 2.7949334027376986e-05,
+ "loss": 1.7699885368347168,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041811846689896,
+ "grad_norm": 0.03982627019286156,
+ "learning_rate": 2.7848865278262427e-05,
+ "loss": 1.741943597793579,
+ "step": 3171
+ },
+ {
+ "epoch": 44.055749128919864,
+ "grad_norm": 0.041303861886262894,
+ "learning_rate": 2.7748618213447268e-05,
+ "loss": 1.7607184648513794,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069686411149824,
+ "grad_norm": 0.04077339917421341,
+ "learning_rate": 2.764859291369885e-05,
+ "loss": 1.7455109357833862,
+ "step": 3173
+ },
+ {
+ "epoch": 44.08362369337979,
+ "grad_norm": 0.04109608381986618,
+ "learning_rate": 2.7548789459605603e-05,
+ "loss": 1.7644455432891846,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09756097560975,
+ "grad_norm": 0.04157571122050285,
+ "learning_rate": 2.744920793157743e-05,
+ "loss": 1.7602050304412842,
+ "step": 3175
+ },
+ {
+ "epoch": 44.11149825783972,
+ "grad_norm": 0.040933798998594284,
+ "learning_rate": 2.7349848409845462e-05,
+ "loss": 1.7585093975067139,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12543554006969,
+ "grad_norm": 0.040963515639305115,
+ "learning_rate": 2.725071097446174e-05,
+ "loss": 1.7687596082687378,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13937282229965,
+ "grad_norm": 0.042252764105796814,
+ "learning_rate": 2.715179570529959e-05,
+ "loss": 1.7694694995880127,
+ "step": 3178
+ },
+ {
+ "epoch": 44.153310104529616,
+ "grad_norm": 0.04096998646855354,
+ "learning_rate": 2.7053102682053134e-05,
+ "loss": 1.7650394439697266,
+ "step": 3179
+ },
+ {
+ "epoch": 44.167247386759584,
+ "grad_norm": 0.04170529916882515,
+ "learning_rate": 2.6954631984237766e-05,
+ "loss": 1.7502206563949585,
+ "step": 3180
+ },
+ {
+ "epoch": 44.181184668989545,
+ "grad_norm": 0.04190446436405182,
+ "learning_rate": 2.6856383691189347e-05,
+ "loss": 1.7572112083435059,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19512195121951,
+ "grad_norm": 0.042252860963344574,
+ "learning_rate": 2.675835788206485e-05,
+ "loss": 1.749457836151123,
+ "step": 3182
+ },
+ {
+ "epoch": 44.20905923344948,
+ "grad_norm": 0.04197750985622406,
+ "learning_rate": 2.6660554635841912e-05,
+ "loss": 1.7559109926223755,
+ "step": 3183
+ },
+ {
+ "epoch": 44.22299651567944,
+ "grad_norm": 0.04213518649339676,
+ "learning_rate": 2.6562974031318788e-05,
+ "loss": 1.7385731935501099,
+ "step": 3184
+ },
+ {
+ "epoch": 44.23693379790941,
+ "grad_norm": 0.04209589958190918,
+ "learning_rate": 2.6465616147114485e-05,
+ "loss": 1.750757098197937,
+ "step": 3185
+ },
+ {
+ "epoch": 44.250871080139376,
+ "grad_norm": 0.04212355986237526,
+ "learning_rate": 2.6368481061668344e-05,
+ "loss": 1.7661514282226562,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26480836236934,
+ "grad_norm": 0.042160581797361374,
+ "learning_rate": 2.6271568853240557e-05,
+ "loss": 1.7580865621566772,
+ "step": 3187
+ },
+ {
+ "epoch": 44.278745644599304,
+ "grad_norm": 0.04218733310699463,
+ "learning_rate": 2.6174879599911347e-05,
+ "loss": 1.7538390159606934,
+ "step": 3188
+ },
+ {
+ "epoch": 44.292682926829265,
+ "grad_norm": 0.04178664833307266,
+ "learning_rate": 2.6078413379581627e-05,
+ "loss": 1.7542293071746826,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30662020905923,
+ "grad_norm": 0.04160580411553383,
+ "learning_rate": 2.5982170269972397e-05,
+ "loss": 1.7468459606170654,
+ "step": 3190
+ },
+ {
+ "epoch": 44.3205574912892,
+ "grad_norm": 0.042538922280073166,
+ "learning_rate": 2.588615034862507e-05,
+ "loss": 1.7589120864868164,
+ "step": 3191
+ },
+ {
+ "epoch": 44.33449477351916,
+ "grad_norm": 0.042292483150959015,
+ "learning_rate": 2.5790353692901158e-05,
+ "loss": 1.759150743484497,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34843205574913,
+ "grad_norm": 0.04133346676826477,
+ "learning_rate": 2.5694780379982238e-05,
+ "loss": 1.757230281829834,
+ "step": 3193
+ },
+ {
+ "epoch": 44.362369337979096,
+ "grad_norm": 0.042723916471004486,
+ "learning_rate": 2.5599430486870055e-05,
+ "loss": 1.7720794677734375,
+ "step": 3194
+ },
+ {
+ "epoch": 44.37630662020906,
+ "grad_norm": 0.04289544001221657,
+ "learning_rate": 2.5504304090386225e-05,
+ "loss": 1.7658957242965698,
+ "step": 3195
+ },
+ {
+ "epoch": 44.390243902439025,
+ "grad_norm": 0.0418098159134388,
+ "learning_rate": 2.5409401267172474e-05,
+ "loss": 1.734413981437683,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40418118466899,
+ "grad_norm": 0.0412493497133255,
+ "learning_rate": 2.5314722093690176e-05,
+ "loss": 1.7341864109039307,
+ "step": 3197
+ },
+ {
+ "epoch": 44.41811846689895,
+ "grad_norm": 0.04258164390921593,
+ "learning_rate": 2.522026664622071e-05,
+ "loss": 1.7577321529388428,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43205574912892,
+ "grad_norm": 0.04071391373872757,
+ "learning_rate": 2.5126035000865124e-05,
+ "loss": 1.7422387599945068,
+ "step": 3199
+ },
+ {
+ "epoch": 44.44599303135889,
+ "grad_norm": 0.04111524671316147,
+ "learning_rate": 2.503202723354413e-05,
+ "loss": 1.7563737630844116,
+ "step": 3200
+ },
+ {
+ "epoch": 44.45993031358885,
+ "grad_norm": 0.0422373004257679,
+ "learning_rate": 2.4938243419998047e-05,
+ "loss": 1.7633488178253174,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47386759581882,
+ "grad_norm": 0.04157081991434097,
+ "learning_rate": 2.484468363578689e-05,
+ "loss": 1.7629499435424805,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48780487804878,
+ "grad_norm": 0.040883488953113556,
+ "learning_rate": 2.475134795629004e-05,
+ "loss": 1.7595467567443848,
+ "step": 3203
+ },
+ {
+ "epoch": 44.501742160278745,
+ "grad_norm": 0.04265756905078888,
+ "learning_rate": 2.4658236456706357e-05,
+ "loss": 1.7728859186172485,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51567944250871,
+ "grad_norm": 0.04229922220110893,
+ "learning_rate": 2.456534921205403e-05,
+ "loss": 1.754500150680542,
+ "step": 3205
+ },
+ {
+ "epoch": 44.52961672473867,
+ "grad_norm": 0.04227421432733536,
+ "learning_rate": 2.4472686297170822e-05,
+ "loss": 1.7667875289916992,
+ "step": 3206
+ },
+ {
+ "epoch": 44.54355400696864,
+ "grad_norm": 0.04271087795495987,
+ "learning_rate": 2.4380247786713403e-05,
+ "loss": 1.7599055767059326,
+ "step": 3207
+ },
+ {
+ "epoch": 44.55749128919861,
+ "grad_norm": 0.04237043857574463,
+ "learning_rate": 2.428803375515788e-05,
+ "loss": 1.74685537815094,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57142857142857,
+ "grad_norm": 0.042595189064741135,
+ "learning_rate": 2.4196044276799397e-05,
+ "loss": 1.7581002712249756,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58536585365854,
+ "grad_norm": 0.04299062490463257,
+ "learning_rate": 2.410427942575228e-05,
+ "loss": 1.757879614830017,
+ "step": 3210
+ },
+ {
+ "epoch": 44.599303135888505,
+ "grad_norm": 0.04270675778388977,
+ "learning_rate": 2.4012739275949746e-05,
+ "loss": 1.7636423110961914,
+ "step": 3211
+ },
+ {
+ "epoch": 44.613240418118465,
+ "grad_norm": 0.042374443262815475,
+ "learning_rate": 2.3921423901144008e-05,
+ "loss": 1.7450188398361206,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62717770034843,
+ "grad_norm": 0.04340631887316704,
+ "learning_rate": 2.3830333374906342e-05,
+ "loss": 1.7522510290145874,
+ "step": 3213
+ },
+ {
+ "epoch": 44.641114982578394,
+ "grad_norm": 0.04277212545275688,
+ "learning_rate": 2.3739467770626637e-05,
+ "loss": 1.7763484716415405,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65505226480836,
+ "grad_norm": 0.04195157438516617,
+ "learning_rate": 2.3648827161513686e-05,
+ "loss": 1.7633893489837646,
+ "step": 3215
+ },
+ {
+ "epoch": 44.66898954703833,
+ "grad_norm": 0.04262831434607506,
+ "learning_rate": 2.3558411620595038e-05,
+ "loss": 1.7620947360992432,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68292682926829,
+ "grad_norm": 0.04257641360163689,
+ "learning_rate": 2.3468221220716844e-05,
+ "loss": 1.757969856262207,
+ "step": 3217
+ },
+ {
+ "epoch": 44.69686411149826,
+ "grad_norm": 0.04366584122180939,
+ "learning_rate": 2.33782560345438e-05,
+ "loss": 1.7606867551803589,
+ "step": 3218
+ },
+ {
+ "epoch": 44.710801393728225,
+ "grad_norm": 0.042488567531108856,
+ "learning_rate": 2.328851613455932e-05,
+ "loss": 1.750715732574463,
+ "step": 3219
+ },
+ {
+ "epoch": 44.724738675958186,
+ "grad_norm": 0.043666984885931015,
+ "learning_rate": 2.3199001593065204e-05,
+ "loss": 1.756069302558899,
+ "step": 3220
+ },
+ {
+ "epoch": 44.73867595818815,
+ "grad_norm": 0.04354606196284294,
+ "learning_rate": 2.3109712482181684e-05,
+ "loss": 1.7556393146514893,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75261324041812,
+ "grad_norm": 0.042452357709407806,
+ "learning_rate": 2.3020648873847414e-05,
+ "loss": 1.7572851181030273,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76655052264808,
+ "grad_norm": 0.043199196457862854,
+ "learning_rate": 2.2931810839819235e-05,
+ "loss": 1.761836290359497,
+ "step": 3223
+ },
+ {
+ "epoch": 44.78048780487805,
+ "grad_norm": 0.04317186400294304,
+ "learning_rate": 2.284319845167253e-05,
+ "loss": 1.7503623962402344,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79442508710802,
+ "grad_norm": 0.04354375973343849,
+ "learning_rate": 2.2754811780800532e-05,
+ "loss": 1.7610375881195068,
+ "step": 3225
+ },
+ {
+ "epoch": 44.80836236933798,
+ "grad_norm": 0.042952246963977814,
+ "learning_rate": 2.2666650898414827e-05,
+ "loss": 1.7500479221343994,
+ "step": 3226
+ },
+ {
+ "epoch": 44.822299651567945,
+ "grad_norm": 0.04257440194487572,
+ "learning_rate": 2.2578715875545095e-05,
+ "loss": 1.7575607299804688,
+ "step": 3227
+ },
+ {
+ "epoch": 44.836236933797906,
+ "grad_norm": 0.041978977620601654,
+ "learning_rate": 2.2491006783038976e-05,
+ "loss": 1.7537864446640015,
+ "step": 3228
+ },
+ {
+ "epoch": 44.850174216027874,
+ "grad_norm": 0.04331148788332939,
+ "learning_rate": 2.2403523691562092e-05,
+ "loss": 1.772139072418213,
+ "step": 3229
+ },
+ {
+ "epoch": 44.86411149825784,
+ "grad_norm": 0.04248177632689476,
+ "learning_rate": 2.2316266671598003e-05,
+ "loss": 1.7485888004302979,
+ "step": 3230
+ },
+ {
+ "epoch": 44.8780487804878,
+ "grad_norm": 0.04298631474375725,
+ "learning_rate": 2.222923579344805e-05,
+ "loss": 1.7410145998001099,
+ "step": 3231
+ },
+ {
+ "epoch": 44.89198606271777,
+ "grad_norm": 0.0432223416864872,
+ "learning_rate": 2.2142431127231502e-05,
+ "loss": 1.755599856376648,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90592334494774,
+ "grad_norm": 0.0427008792757988,
+ "learning_rate": 2.2055852742885322e-05,
+ "loss": 1.764238715171814,
+ "step": 3233
+ },
+ {
+ "epoch": 44.9198606271777,
+ "grad_norm": 0.04341079294681549,
+ "learning_rate": 2.196950071016409e-05,
+ "loss": 1.7583544254302979,
+ "step": 3234
+ },
+ {
+ "epoch": 44.933797909407666,
+ "grad_norm": 0.0435490682721138,
+ "learning_rate": 2.1883375098640134e-05,
+ "loss": 1.7442610263824463,
+ "step": 3235
+ },
+ {
+ "epoch": 44.94773519163763,
+ "grad_norm": 0.042946282774209976,
+ "learning_rate": 2.1797475977703352e-05,
+ "loss": 1.751840591430664,
+ "step": 3236
+ },
+ {
+ "epoch": 44.961672473867594,
+ "grad_norm": 0.0440446138381958,
+ "learning_rate": 2.1711803416560964e-05,
+ "loss": 1.7394684553146362,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97560975609756,
+ "grad_norm": 0.04164465516805649,
+ "learning_rate": 2.16263574842379e-05,
+ "loss": 1.7417004108428955,
+ "step": 3238
+ },
+ {
+ "epoch": 44.98954703832753,
+ "grad_norm": 0.04329351335763931,
+ "learning_rate": 2.15411382495764e-05,
+ "loss": 1.7645809650421143,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.036814916878938675,
+ "learning_rate": 2.1456145781236057e-05,
+ "loss": 1.3324573040008545,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 0.6570897102355957,
+ "eval_runtime": 39.5975,
+ "eval_samples_per_second": 61.671,
+ "eval_steps_per_second": 0.505,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01393728222997,
+ "grad_norm": 0.03884855657815933,
+ "learning_rate": 2.1371380147693752e-05,
+ "loss": 1.7439782619476318,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02787456445993,
+ "grad_norm": 0.038962818682193756,
+ "learning_rate": 2.1286841417243662e-05,
+ "loss": 1.7394144535064697,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041811846689896,
+ "grad_norm": 0.03904305025935173,
+ "learning_rate": 2.1202529657997137e-05,
+ "loss": 1.7592380046844482,
+ "step": 3243
+ },
+ {
+ "epoch": 45.055749128919864,
+ "grad_norm": 0.03950241580605507,
+ "learning_rate": 2.1118444937882538e-05,
+ "loss": 1.7572941780090332,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069686411149824,
+ "grad_norm": 0.03983373939990997,
+ "learning_rate": 2.1034587324645456e-05,
+ "loss": 1.7424907684326172,
+ "step": 3245
+ },
+ {
+ "epoch": 45.08362369337979,
+ "grad_norm": 0.03888066112995148,
+ "learning_rate": 2.095095688584846e-05,
+ "loss": 1.7381600141525269,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09756097560975,
+ "grad_norm": 0.039082713425159454,
+ "learning_rate": 2.086755368887107e-05,
+ "loss": 1.7422431707382202,
+ "step": 3247
+ },
+ {
+ "epoch": 45.11149825783972,
+ "grad_norm": 0.04022567346692085,
+ "learning_rate": 2.0784377800909802e-05,
+ "loss": 1.7446564435958862,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12543554006969,
+ "grad_norm": 0.0401272252202034,
+ "learning_rate": 2.0701429288977847e-05,
+ "loss": 1.745950698852539,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13937282229965,
+ "grad_norm": 0.04019590839743614,
+ "learning_rate": 2.0618708219905455e-05,
+ "loss": 1.7566454410552979,
+ "step": 3250
+ },
+ {
+ "epoch": 45.153310104529616,
+ "grad_norm": 0.03971090540289879,
+ "learning_rate": 2.0536214660339398e-05,
+ "loss": 1.7422828674316406,
+ "step": 3251
+ },
+ {
+ "epoch": 45.167247386759584,
+ "grad_norm": 0.03976941481232643,
+ "learning_rate": 2.0453948676743267e-05,
+ "loss": 1.7474949359893799,
+ "step": 3252
+ },
+ {
+ "epoch": 45.181184668989545,
+ "grad_norm": 0.03973665460944176,
+ "learning_rate": 2.037191033539731e-05,
+ "loss": 1.7500340938568115,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19512195121951,
+ "grad_norm": 0.04038200527429581,
+ "learning_rate": 2.029009970239831e-05,
+ "loss": 1.7580212354660034,
+ "step": 3254
+ },
+ {
+ "epoch": 45.20905923344948,
+ "grad_norm": 0.039865922182798386,
+ "learning_rate": 2.0208516843659678e-05,
+ "loss": 1.7602221965789795,
+ "step": 3255
+ },
+ {
+ "epoch": 45.22299651567944,
+ "grad_norm": 0.039503008127212524,
+ "learning_rate": 2.012716182491114e-05,
+ "loss": 1.7707029581069946,
+ "step": 3256
+ },
+ {
+ "epoch": 45.23693379790941,
+ "grad_norm": 0.03969545289874077,
+ "learning_rate": 2.0046034711699048e-05,
+ "loss": 1.7467467784881592,
+ "step": 3257
+ },
+ {
+ "epoch": 45.250871080139376,
+ "grad_norm": 0.03920473903417587,
+ "learning_rate": 1.9965135569386032e-05,
+ "loss": 1.7424752712249756,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26480836236934,
+ "grad_norm": 0.03914092481136322,
+ "learning_rate": 1.9884464463151077e-05,
+ "loss": 1.7559902667999268,
+ "step": 3259
+ },
+ {
+ "epoch": 45.278745644599304,
+ "grad_norm": 0.040628183633089066,
+ "learning_rate": 1.980402145798933e-05,
+ "loss": 1.7467886209487915,
+ "step": 3260
+ },
+ {
+ "epoch": 45.292682926829265,
+ "grad_norm": 0.040299251675605774,
+ "learning_rate": 1.9723806618712484e-05,
+ "loss": 1.7597342729568481,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30662020905923,
+ "grad_norm": 0.04017207399010658,
+ "learning_rate": 1.9643820009947983e-05,
+ "loss": 1.7490410804748535,
+ "step": 3262
+ },
+ {
+ "epoch": 45.3205574912892,
+ "grad_norm": 0.040886715054512024,
+ "learning_rate": 1.9564061696139687e-05,
+ "loss": 1.7531070709228516,
+ "step": 3263
+ },
+ {
+ "epoch": 45.33449477351916,
+ "grad_norm": 0.04056363180279732,
+ "learning_rate": 1.948453174154744e-05,
+ "loss": 1.7573978900909424,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34843205574913,
+ "grad_norm": 0.04046569764614105,
+ "learning_rate": 1.9405230210247042e-05,
+ "loss": 1.7546555995941162,
+ "step": 3265
+ },
+ {
+ "epoch": 45.362369337979096,
+ "grad_norm": 0.04065052419900894,
+ "learning_rate": 1.9326157166130365e-05,
+ "loss": 1.7649509906768799,
+ "step": 3266
+ },
+ {
+ "epoch": 45.37630662020906,
+ "grad_norm": 0.04096256569027901,
+ "learning_rate": 1.9247312672905017e-05,
+ "loss": 1.760366678237915,
+ "step": 3267
+ },
+ {
+ "epoch": 45.390243902439025,
+ "grad_norm": 0.040638986974954605,
+ "learning_rate": 1.9168696794094758e-05,
+ "loss": 1.74626624584198,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40418118466899,
+ "grad_norm": 0.04070726037025452,
+ "learning_rate": 1.9090309593038812e-05,
+ "loss": 1.7376961708068848,
+ "step": 3269
+ },
+ {
+ "epoch": 45.41811846689895,
+ "grad_norm": 0.040031321346759796,
+ "learning_rate": 1.9012151132892425e-05,
+ "loss": 1.7526248693466187,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43205574912892,
+ "grad_norm": 0.04004877433180809,
+ "learning_rate": 1.893422147662644e-05,
+ "loss": 1.7452869415283203,
+ "step": 3271
+ },
+ {
+ "epoch": 45.44599303135889,
+ "grad_norm": 0.040762972086668015,
+ "learning_rate": 1.8856520687027325e-05,
+ "loss": 1.7534310817718506,
+ "step": 3272
+ },
+ {
+ "epoch": 45.45993031358885,
+ "grad_norm": 0.04094758629798889,
+ "learning_rate": 1.8779048826697318e-05,
+ "loss": 1.764294147491455,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47386759581882,
+ "grad_norm": 0.040894001722335815,
+ "learning_rate": 1.8701805958054e-05,
+ "loss": 1.7491389513015747,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48780487804878,
+ "grad_norm": 0.04058431088924408,
+ "learning_rate": 1.8624792143330544e-05,
+ "loss": 1.7462337017059326,
+ "step": 3275
+ },
+ {
+ "epoch": 45.501742160278745,
+ "grad_norm": 0.04095577076077461,
+ "learning_rate": 1.8548007444575656e-05,
+ "loss": 1.7488327026367188,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51567944250871,
+ "grad_norm": 0.04081057012081146,
+ "learning_rate": 1.8471451923653322e-05,
+ "loss": 1.7610487937927246,
+ "step": 3277
+ },
+ {
+ "epoch": 45.52961672473867,
+ "grad_norm": 0.04027934744954109,
+ "learning_rate": 1.8395125642242987e-05,
+ "loss": 1.7510604858398438,
+ "step": 3278
+ },
+ {
+ "epoch": 45.54355400696864,
+ "grad_norm": 0.04140760749578476,
+ "learning_rate": 1.8319028661839343e-05,
+ "loss": 1.7440571784973145,
+ "step": 3279
+ },
+ {
+ "epoch": 45.55749128919861,
+ "grad_norm": 0.04072791337966919,
+ "learning_rate": 1.8243161043752384e-05,
+ "loss": 1.758582353591919,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57142857142857,
+ "grad_norm": 0.040907423943281174,
+ "learning_rate": 1.8167522849107228e-05,
+ "loss": 1.750474452972412,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58536585365854,
+ "grad_norm": 0.04046669602394104,
+ "learning_rate": 1.809211413884418e-05,
+ "loss": 1.7522833347320557,
+ "step": 3282
+ },
+ {
+ "epoch": 45.599303135888505,
+ "grad_norm": 0.04086044058203697,
+ "learning_rate": 1.8016934973718762e-05,
+ "loss": 1.7532641887664795,
+ "step": 3283
+ },
+ {
+ "epoch": 45.613240418118465,
+ "grad_norm": 0.04121251031756401,
+ "learning_rate": 1.7941985414301403e-05,
+ "loss": 1.7638118267059326,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62717770034843,
+ "grad_norm": 0.041021138429641724,
+ "learning_rate": 1.786726552097766e-05,
+ "loss": 1.7506409883499146,
+ "step": 3285
+ },
+ {
+ "epoch": 45.641114982578394,
+ "grad_norm": 0.040979791432619095,
+ "learning_rate": 1.7792775353947908e-05,
+ "loss": 1.7424489259719849,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65505226480836,
+ "grad_norm": 0.041015319526195526,
+ "learning_rate": 1.7718514973227634e-05,
+ "loss": 1.7470786571502686,
+ "step": 3287
+ },
+ {
+ "epoch": 45.66898954703833,
+ "grad_norm": 0.04148256033658981,
+ "learning_rate": 1.7644484438647057e-05,
+ "loss": 1.7639644145965576,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68292682926829,
+ "grad_norm": 0.041633427143096924,
+ "learning_rate": 1.75706838098512e-05,
+ "loss": 1.75250244140625,
+ "step": 3289
+ },
+ {
+ "epoch": 45.69686411149826,
+ "grad_norm": 0.04139097407460213,
+ "learning_rate": 1.749711314629992e-05,
+ "loss": 1.761831283569336,
+ "step": 3290
+ },
+ {
+ "epoch": 45.710801393728225,
+ "grad_norm": 0.041639216244220734,
+ "learning_rate": 1.742377250726779e-05,
+ "loss": 1.7615015506744385,
+ "step": 3291
+ },
+ {
+ "epoch": 45.724738675958186,
+ "grad_norm": 0.04068813845515251,
+ "learning_rate": 1.7350661951844065e-05,
+ "loss": 1.7496097087860107,
+ "step": 3292
+ },
+ {
+ "epoch": 45.73867595818815,
+ "grad_norm": 0.041004955768585205,
+ "learning_rate": 1.727778153893248e-05,
+ "loss": 1.7539699077606201,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75261324041812,
+ "grad_norm": 0.04117907956242561,
+ "learning_rate": 1.720513132725162e-05,
+ "loss": 1.7560151815414429,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76655052264808,
+ "grad_norm": 0.04089052975177765,
+ "learning_rate": 1.713271137533433e-05,
+ "loss": 1.7342592477798462,
+ "step": 3295
+ },
+ {
+ "epoch": 45.78048780487805,
+ "grad_norm": 0.04100259765982628,
+ "learning_rate": 1.7060521741528127e-05,
+ "loss": 1.7467336654663086,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79442508710802,
+ "grad_norm": 0.04092548415064812,
+ "learning_rate": 1.698856248399482e-05,
+ "loss": 1.7462860345840454,
+ "step": 3297
+ },
+ {
+ "epoch": 45.80836236933798,
+ "grad_norm": 0.0412590429186821,
+ "learning_rate": 1.6916833660710785e-05,
+ "loss": 1.756965160369873,
+ "step": 3298
+ },
+ {
+ "epoch": 45.822299651567945,
+ "grad_norm": 0.041289322078228,
+ "learning_rate": 1.6845335329466606e-05,
+ "loss": 1.7528945207595825,
+ "step": 3299
+ },
+ {
+ "epoch": 45.836236933797906,
+ "grad_norm": 0.041395362466573715,
+ "learning_rate": 1.6774067547867113e-05,
+ "loss": 1.743608832359314,
+ "step": 3300
+ },
+ {
+ "epoch": 45.850174216027874,
+ "grad_norm": 0.04157259687781334,
+ "learning_rate": 1.6703030373331585e-05,
+ "loss": 1.7555323839187622,
+ "step": 3301
+ },
+ {
+ "epoch": 45.86411149825784,
+ "grad_norm": 0.04188670963048935,
+ "learning_rate": 1.6632223863093328e-05,
+ "loss": 1.7543416023254395,
+ "step": 3302
+ },
+ {
+ "epoch": 45.8780487804878,
+ "grad_norm": 0.04152436554431915,
+ "learning_rate": 1.656164807419988e-05,
+ "loss": 1.7428176403045654,
+ "step": 3303
+ },
+ {
+ "epoch": 45.89198606271777,
+ "grad_norm": 0.04157896712422371,
+ "learning_rate": 1.6491303063512866e-05,
+ "loss": 1.7722973823547363,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90592334494774,
+ "grad_norm": 0.04148821160197258,
+ "learning_rate": 1.6421188887708013e-05,
+ "loss": 1.7518692016601562,
+ "step": 3305
+ },
+ {
+ "epoch": 45.9198606271777,
+ "grad_norm": 0.04217620566487312,
+ "learning_rate": 1.6351305603275067e-05,
+ "loss": 1.7555122375488281,
+ "step": 3306
+ },
+ {
+ "epoch": 45.933797909407666,
+ "grad_norm": 0.04238545522093773,
+ "learning_rate": 1.6281653266517658e-05,
+ "loss": 1.7391738891601562,
+ "step": 3307
+ },
+ {
+ "epoch": 45.94773519163763,
+ "grad_norm": 0.0405268594622612,
+ "learning_rate": 1.6212231933553485e-05,
+ "loss": 1.7466099262237549,
+ "step": 3308
+ },
+ {
+ "epoch": 45.961672473867594,
+ "grad_norm": 0.04271813854575157,
+ "learning_rate": 1.6143041660313997e-05,
+ "loss": 1.7705459594726562,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97560975609756,
+ "grad_norm": 0.041771601885557175,
+ "learning_rate": 1.6074082502544668e-05,
+ "loss": 1.740614891052246,
+ "step": 3310
+ },
+ {
+ "epoch": 45.98954703832753,
+ "grad_norm": 0.04159248247742653,
+ "learning_rate": 1.6005354515804503e-05,
+ "loss": 1.7455497980117798,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.03512546420097351,
+ "learning_rate": 1.59368577554664e-05,
+ "loss": 1.3111505508422852,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 0.6581134796142578,
+ "eval_runtime": 40.6459,
+ "eval_samples_per_second": 60.08,
+ "eval_steps_per_second": 0.492,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01393728222997,
+ "grad_norm": 0.038146864622831345,
+ "learning_rate": 1.5868592276717137e-05,
+ "loss": 1.7331663370132446,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02787456445993,
+ "grad_norm": 0.03792983293533325,
+ "learning_rate": 1.5800558134556806e-05,
+ "loss": 1.7528561353683472,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041811846689896,
+ "grad_norm": 0.03715579956769943,
+ "learning_rate": 1.5732755383799315e-05,
+ "loss": 1.7442715167999268,
+ "step": 3315
+ },
+ {
+ "epoch": 46.055749128919864,
+ "grad_norm": 0.0378536619246006,
+ "learning_rate": 1.5665184079072162e-05,
+ "loss": 1.7501194477081299,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069686411149824,
+ "grad_norm": 0.03863883763551712,
+ "learning_rate": 1.5597844274816352e-05,
+ "loss": 1.7525763511657715,
+ "step": 3317
+ },
+ {
+ "epoch": 46.08362369337979,
+ "grad_norm": 0.03858881816267967,
+ "learning_rate": 1.553073602528626e-05,
+ "loss": 1.75484037399292,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09756097560975,
+ "grad_norm": 0.03835397958755493,
+ "learning_rate": 1.546385938454982e-05,
+ "loss": 1.7481768131256104,
+ "step": 3319
+ },
+ {
+ "epoch": 46.11149825783972,
+ "grad_norm": 0.037967924028635025,
+ "learning_rate": 1.5397214406488374e-05,
+ "loss": 1.7442541122436523,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12543554006969,
+ "grad_norm": 0.03828579559922218,
+ "learning_rate": 1.533080114479657e-05,
+ "loss": 1.72530198097229,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13937282229965,
+ "grad_norm": 0.038312915712594986,
+ "learning_rate": 1.5264619652982352e-05,
+ "loss": 1.762467861175537,
+ "step": 3322
+ },
+ {
+ "epoch": 46.153310104529616,
+ "grad_norm": 0.03915497288107872,
+ "learning_rate": 1.5198669984366876e-05,
+ "loss": 1.7523229122161865,
+ "step": 3323
+ },
+ {
+ "epoch": 46.167247386759584,
+ "grad_norm": 0.038662031292915344,
+ "learning_rate": 1.5132952192084793e-05,
+ "loss": 1.7455110549926758,
+ "step": 3324
+ },
+ {
+ "epoch": 46.181184668989545,
+ "grad_norm": 0.03930022567510605,
+ "learning_rate": 1.5067466329083533e-05,
+ "loss": 1.7396240234375,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19512195121951,
+ "grad_norm": 0.03904666006565094,
+ "learning_rate": 1.5002212448123997e-05,
+ "loss": 1.741429328918457,
+ "step": 3326
+ },
+ {
+ "epoch": 46.20905923344948,
+ "grad_norm": 0.03969434276223183,
+ "learning_rate": 1.4937190601780052e-05,
+ "loss": 1.7407313585281372,
+ "step": 3327
+ },
+ {
+ "epoch": 46.22299651567944,
+ "grad_norm": 0.03955148905515671,
+ "learning_rate": 1.4872400842438576e-05,
+ "loss": 1.7501394748687744,
+ "step": 3328
+ },
+ {
+ "epoch": 46.23693379790941,
+ "grad_norm": 0.0389072559773922,
+ "learning_rate": 1.4807843222299557e-05,
+ "loss": 1.756149172782898,
+ "step": 3329
+ },
+ {
+ "epoch": 46.250871080139376,
+ "grad_norm": 0.03908930718898773,
+ "learning_rate": 1.4743517793375814e-05,
+ "loss": 1.7429890632629395,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26480836236934,
+ "grad_norm": 0.038830604404211044,
+ "learning_rate": 1.4679424607493315e-05,
+ "loss": 1.7431840896606445,
+ "step": 3331
+ },
+ {
+ "epoch": 46.278745644599304,
+ "grad_norm": 0.03925040736794472,
+ "learning_rate": 1.4615563716290698e-05,
+ "loss": 1.7441012859344482,
+ "step": 3332
+ },
+ {
+ "epoch": 46.292682926829265,
+ "grad_norm": 0.03930246829986572,
+ "learning_rate": 1.455193517121951e-05,
+ "loss": 1.7456271648406982,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30662020905923,
+ "grad_norm": 0.03882100060582161,
+ "learning_rate": 1.448853902354421e-05,
+ "loss": 1.7365490198135376,
+ "step": 3334
+ },
+ {
+ "epoch": 46.3205574912892,
+ "grad_norm": 0.03973463177680969,
+ "learning_rate": 1.4425375324341827e-05,
+ "loss": 1.7512645721435547,
+ "step": 3335
+ },
+ {
+ "epoch": 46.33449477351916,
+ "grad_norm": 0.039043035358190536,
+ "learning_rate": 1.4362444124502309e-05,
+ "loss": 1.7397332191467285,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34843205574913,
+ "grad_norm": 0.03925980255007744,
+ "learning_rate": 1.429974547472814e-05,
+ "loss": 1.745099425315857,
+ "step": 3337
+ },
+ {
+ "epoch": 46.362369337979096,
+ "grad_norm": 0.03940001502633095,
+ "learning_rate": 1.4237279425534442e-05,
+ "loss": 1.7560515403747559,
+ "step": 3338
+ },
+ {
+ "epoch": 46.37630662020906,
+ "grad_norm": 0.039275459945201874,
+ "learning_rate": 1.4175046027249065e-05,
+ "loss": 1.7552831172943115,
+ "step": 3339
+ },
+ {
+ "epoch": 46.390243902439025,
+ "grad_norm": 0.03995866701006889,
+ "learning_rate": 1.4113045330012324e-05,
+ "loss": 1.7554585933685303,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40418118466899,
+ "grad_norm": 0.04002167284488678,
+ "learning_rate": 1.4051277383776953e-05,
+ "loss": 1.752820611000061,
+ "step": 3341
+ },
+ {
+ "epoch": 46.41811846689895,
+ "grad_norm": 0.04005127772688866,
+ "learning_rate": 1.398974223830839e-05,
+ "loss": 1.7418264150619507,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43205574912892,
+ "grad_norm": 0.03989463299512863,
+ "learning_rate": 1.3928439943184394e-05,
+ "loss": 1.7482575178146362,
+ "step": 3343
+ },
+ {
+ "epoch": 46.44599303135889,
+ "grad_norm": 0.03945402055978775,
+ "learning_rate": 1.3867370547795065e-05,
+ "loss": 1.7270450592041016,
+ "step": 3344
+ },
+ {
+ "epoch": 46.45993031358885,
+ "grad_norm": 0.03910071775317192,
+ "learning_rate": 1.3806534101342891e-05,
+ "loss": 1.7461726665496826,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47386759581882,
+ "grad_norm": 0.040124330669641495,
+ "learning_rate": 1.3745930652842728e-05,
+ "loss": 1.7472071647644043,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48780487804878,
+ "grad_norm": 0.03996257483959198,
+ "learning_rate": 1.368556025112176e-05,
+ "loss": 1.7534830570220947,
+ "step": 3347
+ },
+ {
+ "epoch": 46.501742160278745,
+ "grad_norm": 0.039497364312410355,
+ "learning_rate": 1.362542294481918e-05,
+ "loss": 1.7417938709259033,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51567944250871,
+ "grad_norm": 0.040586575865745544,
+ "learning_rate": 1.356551878238656e-05,
+ "loss": 1.7480063438415527,
+ "step": 3349
+ },
+ {
+ "epoch": 46.52961672473867,
+ "grad_norm": 0.03976017236709595,
+ "learning_rate": 1.3505847812087724e-05,
+ "loss": 1.7456187009811401,
+ "step": 3350
+ },
+ {
+ "epoch": 46.54355400696864,
+ "grad_norm": 0.03951919451355934,
+ "learning_rate": 1.3446410081998384e-05,
+ "loss": 1.7342455387115479,
+ "step": 3351
+ },
+ {
+ "epoch": 46.55749128919861,
+ "grad_norm": 0.04001230373978615,
+ "learning_rate": 1.3387205640006494e-05,
+ "loss": 1.7555949687957764,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57142857142857,
+ "grad_norm": 0.03985046222805977,
+ "learning_rate": 1.332823453381196e-05,
+ "loss": 1.7450182437896729,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58536585365854,
+ "grad_norm": 0.04045078530907631,
+ "learning_rate": 1.3269496810926807e-05,
+ "loss": 1.7479677200317383,
+ "step": 3354
+ },
+ {
+ "epoch": 46.599303135888505,
+ "grad_norm": 0.04043098911643028,
+ "learning_rate": 1.3210992518674848e-05,
+ "loss": 1.7537192106246948,
+ "step": 3355
+ },
+ {
+ "epoch": 46.613240418118465,
+ "grad_norm": 0.03941028565168381,
+ "learning_rate": 1.315272170419201e-05,
+ "loss": 1.7403452396392822,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62717770034843,
+ "grad_norm": 0.04040810465812683,
+ "learning_rate": 1.3094684414426073e-05,
+ "loss": 1.7479832172393799,
+ "step": 3357
+ },
+ {
+ "epoch": 46.641114982578394,
+ "grad_norm": 0.0394478514790535,
+ "learning_rate": 1.3036880696136532e-05,
+ "loss": 1.7441142797470093,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65505226480836,
+ "grad_norm": 0.040056221187114716,
+ "learning_rate": 1.2979310595894882e-05,
+ "loss": 1.7515372037887573,
+ "step": 3359
+ },
+ {
+ "epoch": 46.66898954703833,
+ "grad_norm": 0.03989299759268761,
+ "learning_rate": 1.2921974160084251e-05,
+ "loss": 1.7415657043457031,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68292682926829,
+ "grad_norm": 0.0411345548927784,
+ "learning_rate": 1.2864871434899672e-05,
+ "loss": 1.7447781562805176,
+ "step": 3361
+ },
+ {
+ "epoch": 46.69686411149826,
+ "grad_norm": 0.039821773767471313,
+ "learning_rate": 1.2808002466347654e-05,
+ "loss": 1.7494394779205322,
+ "step": 3362
+ },
+ {
+ "epoch": 46.710801393728225,
+ "grad_norm": 0.04027014598250389,
+ "learning_rate": 1.2751367300246535e-05,
+ "loss": 1.760572910308838,
+ "step": 3363
+ },
+ {
+ "epoch": 46.724738675958186,
+ "grad_norm": 0.04048950597643852,
+ "learning_rate": 1.2694965982226301e-05,
+ "loss": 1.7438969612121582,
+ "step": 3364
+ },
+ {
+ "epoch": 46.73867595818815,
+ "grad_norm": 0.03991667553782463,
+ "learning_rate": 1.2638798557728402e-05,
+ "loss": 1.7614572048187256,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75261324041812,
+ "grad_norm": 0.04084254056215286,
+ "learning_rate": 1.2582865072006e-05,
+ "loss": 1.7515487670898438,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76655052264808,
+ "grad_norm": 0.039985332638025284,
+ "learning_rate": 1.2527165570123586e-05,
+ "loss": 1.7455508708953857,
+ "step": 3367
+ },
+ {
+ "epoch": 46.78048780487805,
+ "grad_norm": 0.04038294404745102,
+ "learning_rate": 1.2471700096957376e-05,
+ "loss": 1.7448019981384277,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79442508710802,
+ "grad_norm": 0.039845626801252365,
+ "learning_rate": 1.2416468697194767e-05,
+ "loss": 1.7415070533752441,
+ "step": 3369
+ },
+ {
+ "epoch": 46.80836236933798,
+ "grad_norm": 0.040511518716812134,
+ "learning_rate": 1.2361471415334795e-05,
+ "loss": 1.7448599338531494,
+ "step": 3370
+ },
+ {
+ "epoch": 46.822299651567945,
+ "grad_norm": 0.040095821022987366,
+ "learning_rate": 1.2306708295687706e-05,
+ "loss": 1.7481844425201416,
+ "step": 3371
+ },
+ {
+ "epoch": 46.836236933797906,
+ "grad_norm": 0.03934630751609802,
+ "learning_rate": 1.2252179382375211e-05,
+ "loss": 1.7475495338439941,
+ "step": 3372
+ },
+ {
+ "epoch": 46.850174216027874,
+ "grad_norm": 0.039827462285757065,
+ "learning_rate": 1.2197884719330281e-05,
+ "loss": 1.7375001907348633,
+ "step": 3373
+ },
+ {
+ "epoch": 46.86411149825784,
+ "grad_norm": 0.039957933127880096,
+ "learning_rate": 1.2143824350297104e-05,
+ "loss": 1.7417705059051514,
+ "step": 3374
+ },
+ {
+ "epoch": 46.8780487804878,
+ "grad_norm": 0.04053289443254471,
+ "learning_rate": 1.2089998318831137e-05,
+ "loss": 1.7526462078094482,
+ "step": 3375
+ },
+ {
+ "epoch": 46.89198606271777,
+ "grad_norm": 0.03925419971346855,
+ "learning_rate": 1.2036406668299062e-05,
+ "loss": 1.746978998184204,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90592334494774,
+ "grad_norm": 0.040124308317899704,
+ "learning_rate": 1.198304944187868e-05,
+ "loss": 1.7493703365325928,
+ "step": 3377
+ },
+ {
+ "epoch": 46.9198606271777,
+ "grad_norm": 0.04053357243537903,
+ "learning_rate": 1.1929926682558995e-05,
+ "loss": 1.7561579942703247,
+ "step": 3378
+ },
+ {
+ "epoch": 46.933797909407666,
+ "grad_norm": 0.040585730224847794,
+ "learning_rate": 1.1877038433139998e-05,
+ "loss": 1.7525160312652588,
+ "step": 3379
+ },
+ {
+ "epoch": 46.94773519163763,
+ "grad_norm": 0.040667396038770676,
+ "learning_rate": 1.1824384736232879e-05,
+ "loss": 1.7589459419250488,
+ "step": 3380
+ },
+ {
+ "epoch": 46.961672473867594,
+ "grad_norm": 0.039576202630996704,
+ "learning_rate": 1.1771965634259653e-05,
+ "loss": 1.771327257156372,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97560975609756,
+ "grad_norm": 0.04068684205412865,
+ "learning_rate": 1.1719781169453536e-05,
+ "loss": 1.748750925064087,
+ "step": 3382
+ },
+ {
+ "epoch": 46.98954703832753,
+ "grad_norm": 0.04068964719772339,
+ "learning_rate": 1.1667831383858562e-05,
+ "loss": 1.7589222192764282,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.03399408236145973,
+ "learning_rate": 1.1616116319329808e-05,
+ "loss": 1.3100368976593018,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 0.6589331030845642,
+ "eval_runtime": 40.2755,
+ "eval_samples_per_second": 60.632,
+ "eval_steps_per_second": 0.497,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01393728222997,
+ "grad_norm": 0.03671970218420029,
+ "learning_rate": 1.1564636017533036e-05,
+ "loss": 1.7442817687988281,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02787456445993,
+ "grad_norm": 0.036766838282346725,
+ "learning_rate": 1.151339051994508e-05,
+ "loss": 1.7399423122406006,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041811846689896,
+ "grad_norm": 0.037153277546167374,
+ "learning_rate": 1.146237986785356e-05,
+ "loss": 1.7389707565307617,
+ "step": 3387
+ },
+ {
+ "epoch": 47.055749128919864,
+ "grad_norm": 0.03724278137087822,
+ "learning_rate": 1.141160410235674e-05,
+ "loss": 1.7309342622756958,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069686411149824,
+ "grad_norm": 0.03884275257587433,
+ "learning_rate": 1.1361063264363776e-05,
+ "loss": 1.7472870349884033,
+ "step": 3389
+ },
+ {
+ "epoch": 47.08362369337979,
+ "grad_norm": 0.03773004934191704,
+ "learning_rate": 1.1310757394594514e-05,
+ "loss": 1.748953938484192,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09756097560975,
+ "grad_norm": 0.03674894943833351,
+ "learning_rate": 1.1260686533579481e-05,
+ "loss": 1.7510933876037598,
+ "step": 3391
+ },
+ {
+ "epoch": 47.11149825783972,
+ "grad_norm": 0.038440484553575516,
+ "learning_rate": 1.121085072165986e-05,
+ "loss": 1.734560489654541,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12543554006969,
+ "grad_norm": 0.03765211999416351,
+ "learning_rate": 1.1161249998987429e-05,
+ "loss": 1.742401123046875,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13937282229965,
+ "grad_norm": 0.037882134318351746,
+ "learning_rate": 1.111188440552471e-05,
+ "loss": 1.750654697418213,
+ "step": 3394
+ },
+ {
+ "epoch": 47.153310104529616,
+ "grad_norm": 0.038071002811193466,
+ "learning_rate": 1.106275398104455e-05,
+ "loss": 1.7491934299468994,
+ "step": 3395
+ },
+ {
+ "epoch": 47.167247386759584,
+ "grad_norm": 0.037732210010290146,
+ "learning_rate": 1.1013858765130468e-05,
+ "loss": 1.7402474880218506,
+ "step": 3396
+ },
+ {
+ "epoch": 47.181184668989545,
+ "grad_norm": 0.03905113786458969,
+ "learning_rate": 1.0965198797176513e-05,
+ "loss": 1.7507375478744507,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19512195121951,
+ "grad_norm": 0.038533765822649,
+ "learning_rate": 1.0916774116387103e-05,
+ "loss": 1.7438175678253174,
+ "step": 3398
+ },
+ {
+ "epoch": 47.20905923344948,
+ "grad_norm": 0.03780142962932587,
+ "learning_rate": 1.0868584761777125e-05,
+ "loss": 1.7474523782730103,
+ "step": 3399
+ },
+ {
+ "epoch": 47.22299651567944,
+ "grad_norm": 0.03854325786232948,
+ "learning_rate": 1.0820630772171869e-05,
+ "loss": 1.7482647895812988,
+ "step": 3400
+ },
+ {
+ "epoch": 47.23693379790941,
+ "grad_norm": 0.04027657210826874,
+ "learning_rate": 1.0772912186206987e-05,
+ "loss": 1.7588740587234497,
+ "step": 3401
+ },
+ {
+ "epoch": 47.250871080139376,
+ "grad_norm": 0.03937956318259239,
+ "learning_rate": 1.0725429042328542e-05,
+ "loss": 1.7503528594970703,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26480836236934,
+ "grad_norm": 0.03844330832362175,
+ "learning_rate": 1.0678181378792794e-05,
+ "loss": 1.7492938041687012,
+ "step": 3403
+ },
+ {
+ "epoch": 47.278745644599304,
+ "grad_norm": 0.038146719336509705,
+ "learning_rate": 1.0631169233666306e-05,
+ "loss": 1.7423465251922607,
+ "step": 3404
+ },
+ {
+ "epoch": 47.292682926829265,
+ "grad_norm": 0.039059530943632126,
+ "learning_rate": 1.0584392644826016e-05,
+ "loss": 1.747842788696289,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30662020905923,
+ "grad_norm": 0.03757455572485924,
+ "learning_rate": 1.053785164995886e-05,
+ "loss": 1.7496275901794434,
+ "step": 3406
+ },
+ {
+ "epoch": 47.3205574912892,
+ "grad_norm": 0.037678398191928864,
+ "learning_rate": 1.0491546286562148e-05,
+ "loss": 1.7476543188095093,
+ "step": 3407
+ },
+ {
+ "epoch": 47.33449477351916,
+ "grad_norm": 0.03855350241065025,
+ "learning_rate": 1.0445476591943259e-05,
+ "loss": 1.740640640258789,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34843205574913,
+ "grad_norm": 0.03885442763566971,
+ "learning_rate": 1.0399642603219712e-05,
+ "loss": 1.7487856149673462,
+ "step": 3409
+ },
+ {
+ "epoch": 47.362369337979096,
+ "grad_norm": 0.03796287998557091,
+ "learning_rate": 1.0354044357319129e-05,
+ "loss": 1.7372463941574097,
+ "step": 3410
+ },
+ {
+ "epoch": 47.37630662020906,
+ "grad_norm": 0.038020867854356766,
+ "learning_rate": 1.0308681890979173e-05,
+ "loss": 1.7441340684890747,
+ "step": 3411
+ },
+ {
+ "epoch": 47.390243902439025,
+ "grad_norm": 0.03857402503490448,
+ "learning_rate": 1.026355524074758e-05,
+ "loss": 1.746065378189087,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40418118466899,
+ "grad_norm": 0.03823517635464668,
+ "learning_rate": 1.0218664442982084e-05,
+ "loss": 1.7378289699554443,
+ "step": 3413
+ },
+ {
+ "epoch": 47.41811846689895,
+ "grad_norm": 0.038548193871974945,
+ "learning_rate": 1.0174009533850368e-05,
+ "loss": 1.7585103511810303,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43205574912892,
+ "grad_norm": 0.03859351947903633,
+ "learning_rate": 1.0129590549330119e-05,
+ "loss": 1.7489036321640015,
+ "step": 3415
+ },
+ {
+ "epoch": 47.44599303135889,
+ "grad_norm": 0.03980541229248047,
+ "learning_rate": 1.0085407525208865e-05,
+ "loss": 1.7471911907196045,
+ "step": 3416
+ },
+ {
+ "epoch": 47.45993031358885,
+ "grad_norm": 0.03822151571512222,
+ "learning_rate": 1.0041460497084135e-05,
+ "loss": 1.7424052953720093,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47386759581882,
+ "grad_norm": 0.03830152377486229,
+ "learning_rate": 9.997749500363172e-06,
+ "loss": 1.7360411882400513,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48780487804878,
+ "grad_norm": 0.03838806599378586,
+ "learning_rate": 9.954274570263191e-06,
+ "loss": 1.7423158884048462,
+ "step": 3419
+ },
+ {
+ "epoch": 47.501742160278745,
+ "grad_norm": 0.03949080407619476,
+ "learning_rate": 9.911035741811153e-06,
+ "loss": 1.7440277338027954,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51567944250871,
+ "grad_norm": 0.03809436038136482,
+ "learning_rate": 9.868033049843787e-06,
+ "loss": 1.7273075580596924,
+ "step": 3421
+ },
+ {
+ "epoch": 47.52961672473867,
+ "grad_norm": 0.03868911787867546,
+ "learning_rate": 9.825266529007601e-06,
+ "loss": 1.750636339187622,
+ "step": 3422
+ },
+ {
+ "epoch": 47.54355400696864,
+ "grad_norm": 0.03874430060386658,
+ "learning_rate": 9.782736213758782e-06,
+ "loss": 1.764697551727295,
+ "step": 3423
+ },
+ {
+ "epoch": 47.55749128919861,
+ "grad_norm": 0.03880039602518082,
+ "learning_rate": 9.740442138363317e-06,
+ "loss": 1.7470829486846924,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57142857142857,
+ "grad_norm": 0.03893805295228958,
+ "learning_rate": 9.698384336896679e-06,
+ "loss": 1.743140697479248,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58536585365854,
+ "grad_norm": 0.03870455175638199,
+ "learning_rate": 9.656562843244114e-06,
+ "loss": 1.7481999397277832,
+ "step": 3426
+ },
+ {
+ "epoch": 47.599303135888505,
+ "grad_norm": 0.038351476192474365,
+ "learning_rate": 9.614977691100445e-06,
+ "loss": 1.7535104751586914,
+ "step": 3427
+ },
+ {
+ "epoch": 47.613240418118465,
+ "grad_norm": 0.03865189477801323,
+ "learning_rate": 9.57362891397007e-06,
+ "loss": 1.7472349405288696,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62717770034843,
+ "grad_norm": 0.03885399177670479,
+ "learning_rate": 9.532516545167004e-06,
+ "loss": 1.74338698387146,
+ "step": 3429
+ },
+ {
+ "epoch": 47.641114982578394,
+ "grad_norm": 0.03872792050242424,
+ "learning_rate": 9.491640617814634e-06,
+ "loss": 1.7531355619430542,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65505226480836,
+ "grad_norm": 0.039299190044403076,
+ "learning_rate": 9.451001164846092e-06,
+ "loss": 1.751693606376648,
+ "step": 3431
+ },
+ {
+ "epoch": 47.66898954703833,
+ "grad_norm": 0.03835092857480049,
+ "learning_rate": 9.410598219003725e-06,
+ "loss": 1.740051031112671,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68292682926829,
+ "grad_norm": 0.0389421246945858,
+ "learning_rate": 9.370431812839556e-06,
+ "loss": 1.739893913269043,
+ "step": 3433
+ },
+ {
+ "epoch": 47.69686411149826,
+ "grad_norm": 0.037790242582559586,
+ "learning_rate": 9.330501978714857e-06,
+ "loss": 1.7449544668197632,
+ "step": 3434
+ },
+ {
+ "epoch": 47.710801393728225,
+ "grad_norm": 0.038736164569854736,
+ "learning_rate": 9.290808748800408e-06,
+ "loss": 1.7493153810501099,
+ "step": 3435
+ },
+ {
+ "epoch": 47.724738675958186,
+ "grad_norm": 0.038166582584381104,
+ "learning_rate": 9.25135215507634e-06,
+ "loss": 1.7426971197128296,
+ "step": 3436
+ },
+ {
+ "epoch": 47.73867595818815,
+ "grad_norm": 0.038498617708683014,
+ "learning_rate": 9.212132229332094e-06,
+ "loss": 1.731830358505249,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75261324041812,
+ "grad_norm": 0.038688234984874725,
+ "learning_rate": 9.173149003166458e-06,
+ "loss": 1.7332818508148193,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76655052264808,
+ "grad_norm": 0.03849519416689873,
+ "learning_rate": 9.134402507987533e-06,
+ "loss": 1.7459259033203125,
+ "step": 3439
+ },
+ {
+ "epoch": 47.78048780487805,
+ "grad_norm": 0.03904317691922188,
+ "learning_rate": 9.095892775012672e-06,
+ "loss": 1.740927815437317,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79442508710802,
+ "grad_norm": 0.03893847018480301,
+ "learning_rate": 9.057619835268437e-06,
+ "loss": 1.7319128513336182,
+ "step": 3441
+ },
+ {
+ "epoch": 47.80836236933798,
+ "grad_norm": 0.03848448768258095,
+ "learning_rate": 9.019583719590641e-06,
+ "loss": 1.7380175590515137,
+ "step": 3442
+ },
+ {
+ "epoch": 47.822299651567945,
+ "grad_norm": 0.0387788787484169,
+ "learning_rate": 8.981784458624375e-06,
+ "loss": 1.735992193222046,
+ "step": 3443
+ },
+ {
+ "epoch": 47.836236933797906,
+ "grad_norm": 0.038397762924432755,
+ "learning_rate": 8.944222082823717e-06,
+ "loss": 1.7542471885681152,
+ "step": 3444
+ },
+ {
+ "epoch": 47.850174216027874,
+ "grad_norm": 0.03903001919388771,
+ "learning_rate": 8.906896622452055e-06,
+ "loss": 1.7547940015792847,
+ "step": 3445
+ },
+ {
+ "epoch": 47.86411149825784,
+ "grad_norm": 0.039537038654088974,
+ "learning_rate": 8.86980810758183e-06,
+ "loss": 1.7347489595413208,
+ "step": 3446
+ },
+ {
+ "epoch": 47.8780487804878,
+ "grad_norm": 0.03893627971410751,
+ "learning_rate": 8.832956568094596e-06,
+ "loss": 1.746551513671875,
+ "step": 3447
+ },
+ {
+ "epoch": 47.89198606271777,
+ "grad_norm": 0.03855917230248451,
+ "learning_rate": 8.796342033680928e-06,
+ "loss": 1.7296223640441895,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90592334494774,
+ "grad_norm": 0.03921154886484146,
+ "learning_rate": 8.759964533840574e-06,
+ "loss": 1.7452442646026611,
+ "step": 3449
+ },
+ {
+ "epoch": 47.9198606271777,
+ "grad_norm": 0.038231946527957916,
+ "learning_rate": 8.723824097882145e-06,
+ "loss": 1.7400989532470703,
+ "step": 3450
+ },
+ {
+ "epoch": 47.933797909407666,
+ "grad_norm": 0.03857921063899994,
+ "learning_rate": 8.687920754923393e-06,
+ "loss": 1.742523431777954,
+ "step": 3451
+ },
+ {
+ "epoch": 47.94773519163763,
+ "grad_norm": 0.039119452238082886,
+ "learning_rate": 8.652254533890956e-06,
+ "loss": 1.745814323425293,
+ "step": 3452
+ },
+ {
+ "epoch": 47.961672473867594,
+ "grad_norm": 0.039589304476976395,
+ "learning_rate": 8.616825463520493e-06,
+ "loss": 1.7558345794677734,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97560975609756,
+ "grad_norm": 0.03909674659371376,
+ "learning_rate": 8.581633572356576e-06,
+ "loss": 1.7320977449417114,
+ "step": 3454
+ },
+ {
+ "epoch": 47.98954703832753,
+ "grad_norm": 0.0385885164141655,
+ "learning_rate": 8.546678888752629e-06,
+ "loss": 1.744365930557251,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.032989442348480225,
+ "learning_rate": 8.51196144087103e-06,
+ "loss": 1.306262493133545,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 0.659737765789032,
+ "eval_runtime": 40.3038,
+ "eval_samples_per_second": 60.59,
+ "eval_steps_per_second": 0.496,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01393728222997,
+ "grad_norm": 0.035367924720048904,
+ "learning_rate": 8.477481256683005e-06,
+ "loss": 1.719075322151184,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02787456445993,
+ "grad_norm": 0.03542623668909073,
+ "learning_rate": 8.44323836396864e-06,
+ "loss": 1.742891788482666,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041811846689896,
+ "grad_norm": 0.03636869788169861,
+ "learning_rate": 8.409232790316763e-06,
+ "loss": 1.7473716735839844,
+ "step": 3459
+ },
+ {
+ "epoch": 48.055749128919864,
+ "grad_norm": 0.03710813447833061,
+ "learning_rate": 8.375464563125094e-06,
+ "loss": 1.7469065189361572,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069686411149824,
+ "grad_norm": 0.03639912232756615,
+ "learning_rate": 8.341933709600102e-06,
+ "loss": 1.750680685043335,
+ "step": 3461
+ },
+ {
+ "epoch": 48.08362369337979,
+ "grad_norm": 0.03568524494767189,
+ "learning_rate": 8.308640256756916e-06,
+ "loss": 1.7470159530639648,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09756097560975,
+ "grad_norm": 0.03604842349886894,
+ "learning_rate": 8.275584231419512e-06,
+ "loss": 1.7345272302627563,
+ "step": 3463
+ },
+ {
+ "epoch": 48.11149825783972,
+ "grad_norm": 0.03659019246697426,
+ "learning_rate": 8.24276566022055e-06,
+ "loss": 1.7318280935287476,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12543554006969,
+ "grad_norm": 0.03665720298886299,
+ "learning_rate": 8.21018456960132e-06,
+ "loss": 1.7370295524597168,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13937282229965,
+ "grad_norm": 0.03737010806798935,
+ "learning_rate": 8.177840985811858e-06,
+ "loss": 1.7536876201629639,
+ "step": 3466
+ },
+ {
+ "epoch": 48.153310104529616,
+ "grad_norm": 0.03616785630583763,
+ "learning_rate": 8.145734934910757e-06,
+ "loss": 1.7425014972686768,
+ "step": 3467
+ },
+ {
+ "epoch": 48.167247386759584,
+ "grad_norm": 0.0366208590567112,
+ "learning_rate": 8.113866442765364e-06,
+ "loss": 1.7445616722106934,
+ "step": 3468
+ },
+ {
+ "epoch": 48.181184668989545,
+ "grad_norm": 0.03636297211050987,
+ "learning_rate": 8.082235535051449e-06,
+ "loss": 1.7424728870391846,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19512195121951,
+ "grad_norm": 0.037299733608961105,
+ "learning_rate": 8.050842237253533e-06,
+ "loss": 1.7560949325561523,
+ "step": 3470
+ },
+ {
+ "epoch": 48.20905923344948,
+ "grad_norm": 0.03679182380437851,
+ "learning_rate": 8.019686574664564e-06,
+ "loss": 1.7488621473312378,
+ "step": 3471
+ },
+ {
+ "epoch": 48.22299651567944,
+ "grad_norm": 0.03720501810312271,
+ "learning_rate": 7.988768572386171e-06,
+ "loss": 1.7504916191101074,
+ "step": 3472
+ },
+ {
+ "epoch": 48.23693379790941,
+ "grad_norm": 0.037201616913080215,
+ "learning_rate": 7.958088255328415e-06,
+ "loss": 1.7474762201309204,
+ "step": 3473
+ },
+ {
+ "epoch": 48.250871080139376,
+ "grad_norm": 0.0363336056470871,
+ "learning_rate": 7.927645648209837e-06,
+ "loss": 1.737074613571167,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26480836236934,
+ "grad_norm": 0.03655460849404335,
+ "learning_rate": 7.897440775557506e-06,
+ "loss": 1.7407782077789307,
+ "step": 3475
+ },
+ {
+ "epoch": 48.278745644599304,
+ "grad_norm": 0.0372798889875412,
+ "learning_rate": 7.867473661706946e-06,
+ "loss": 1.7514898777008057,
+ "step": 3476
+ },
+ {
+ "epoch": 48.292682926829265,
+ "grad_norm": 0.03731389716267586,
+ "learning_rate": 7.837744330802136e-06,
+ "loss": 1.7415683269500732,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30662020905923,
+ "grad_norm": 0.037008073180913925,
+ "learning_rate": 7.808252806795446e-06,
+ "loss": 1.7323089838027954,
+ "step": 3478
+ },
+ {
+ "epoch": 48.3205574912892,
+ "grad_norm": 0.036683958023786545,
+ "learning_rate": 7.778999113447673e-06,
+ "loss": 1.7389119863510132,
+ "step": 3479
+ },
+ {
+ "epoch": 48.33449477351916,
+ "grad_norm": 0.03731265664100647,
+ "learning_rate": 7.749983274327994e-06,
+ "loss": 1.7375361919403076,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34843205574913,
+ "grad_norm": 0.037563782185316086,
+ "learning_rate": 7.721205312813921e-06,
+ "loss": 1.744289517402649,
+ "step": 3481
+ },
+ {
+ "epoch": 48.362369337979096,
+ "grad_norm": 0.03690405562520027,
+ "learning_rate": 7.69266525209132e-06,
+ "loss": 1.736900806427002,
+ "step": 3482
+ },
+ {
+ "epoch": 48.37630662020906,
+ "grad_norm": 0.036237478256225586,
+ "learning_rate": 7.664363115154476e-06,
+ "loss": 1.7253262996673584,
+ "step": 3483
+ },
+ {
+ "epoch": 48.390243902439025,
+ "grad_norm": 0.037100717425346375,
+ "learning_rate": 7.636298924805872e-06,
+ "loss": 1.730082392692566,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40418118466899,
+ "grad_norm": 0.036637891083955765,
+ "learning_rate": 7.608472703656282e-06,
+ "loss": 1.749288558959961,
+ "step": 3485
+ },
+ {
+ "epoch": 48.41811846689895,
+ "grad_norm": 0.03698175400495529,
+ "learning_rate": 7.580884474124833e-06,
+ "loss": 1.7454564571380615,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43205574912892,
+ "grad_norm": 0.03757963329553604,
+ "learning_rate": 7.55353425843888e-06,
+ "loss": 1.739332675933838,
+ "step": 3487
+ },
+ {
+ "epoch": 48.44599303135889,
+ "grad_norm": 0.036970723420381546,
+ "learning_rate": 7.526422078633939e-06,
+ "loss": 1.7400529384613037,
+ "step": 3488
+ },
+ {
+ "epoch": 48.45993031358885,
+ "grad_norm": 0.037274643778800964,
+ "learning_rate": 7.4995479565538476e-06,
+ "loss": 1.7452963590621948,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47386759581882,
+ "grad_norm": 0.03724444657564163,
+ "learning_rate": 7.4729119138506044e-06,
+ "loss": 1.7532882690429688,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48780487804878,
+ "grad_norm": 0.03693290427327156,
+ "learning_rate": 7.446513971984366e-06,
+ "loss": 1.7261027097702026,
+ "step": 3491
+ },
+ {
+ "epoch": 48.501742160278745,
+ "grad_norm": 0.037947945296764374,
+ "learning_rate": 7.420354152223485e-06,
+ "loss": 1.7563652992248535,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51567944250871,
+ "grad_norm": 0.03800833597779274,
+ "learning_rate": 7.394432475644402e-06,
+ "loss": 1.7506392002105713,
+ "step": 3493
+ },
+ {
+ "epoch": 48.52961672473867,
+ "grad_norm": 0.03641824796795845,
+ "learning_rate": 7.368748963131854e-06,
+ "loss": 1.735479712486267,
+ "step": 3494
+ },
+ {
+ "epoch": 48.54355400696864,
+ "grad_norm": 0.038214061409235,
+ "learning_rate": 7.343303635378537e-06,
+ "loss": 1.7464430332183838,
+ "step": 3495
+ },
+ {
+ "epoch": 48.55749128919861,
+ "grad_norm": 0.037863098084926605,
+ "learning_rate": 7.31809651288524e-06,
+ "loss": 1.7333879470825195,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57142857142857,
+ "grad_norm": 0.037542443722486496,
+ "learning_rate": 7.2931276159609475e-06,
+ "loss": 1.7467310428619385,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58536585365854,
+ "grad_norm": 0.03775721788406372,
+ "learning_rate": 7.268396964722602e-06,
+ "loss": 1.7356815338134766,
+ "step": 3498
+ },
+ {
+ "epoch": 48.599303135888505,
+ "grad_norm": 0.03697579726576805,
+ "learning_rate": 7.243904579095308e-06,
+ "loss": 1.732956886291504,
+ "step": 3499
+ },
+ {
+ "epoch": 48.613240418118465,
+ "grad_norm": 0.03738914057612419,
+ "learning_rate": 7.219650478812065e-06,
+ "loss": 1.7363367080688477,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62717770034843,
+ "grad_norm": 0.03715483844280243,
+ "learning_rate": 7.195634683413996e-06,
+ "loss": 1.7278720140457153,
+ "step": 3501
+ },
+ {
+ "epoch": 48.641114982578394,
+ "grad_norm": 0.03724132478237152,
+ "learning_rate": 7.171857212250225e-06,
+ "loss": 1.753392219543457,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65505226480836,
+ "grad_norm": 0.037609368562698364,
+ "learning_rate": 7.148318084477766e-06,
+ "loss": 1.7377769947052002,
+ "step": 3503
+ },
+ {
+ "epoch": 48.66898954703833,
+ "grad_norm": 0.03773123770952225,
+ "learning_rate": 7.125017319061728e-06,
+ "loss": 1.7462949752807617,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68292682926829,
+ "grad_norm": 0.03815515339374542,
+ "learning_rate": 7.101954934775084e-06,
+ "loss": 1.7362608909606934,
+ "step": 3505
+ },
+ {
+ "epoch": 48.69686411149826,
+ "grad_norm": 0.03709382191300392,
+ "learning_rate": 7.079130950198802e-06,
+ "loss": 1.7472469806671143,
+ "step": 3506
+ },
+ {
+ "epoch": 48.710801393728225,
+ "grad_norm": 0.03823981434106827,
+ "learning_rate": 7.05654538372171e-06,
+ "loss": 1.7437620162963867,
+ "step": 3507
+ },
+ {
+ "epoch": 48.724738675958186,
+ "grad_norm": 0.037286169826984406,
+ "learning_rate": 7.034198253540665e-06,
+ "loss": 1.7460813522338867,
+ "step": 3508
+ },
+ {
+ "epoch": 48.73867595818815,
+ "grad_norm": 0.03855825960636139,
+ "learning_rate": 7.012089577660255e-06,
+ "loss": 1.7547013759613037,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75261324041812,
+ "grad_norm": 0.038815904408693314,
+ "learning_rate": 6.990219373893128e-06,
+ "loss": 1.7504229545593262,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76655052264808,
+ "grad_norm": 0.03781234472990036,
+ "learning_rate": 6.968587659859665e-06,
+ "loss": 1.7378299236297607,
+ "step": 3511
+ },
+ {
+ "epoch": 48.78048780487805,
+ "grad_norm": 0.036778468638658524,
+ "learning_rate": 6.9471944529881376e-06,
+ "loss": 1.7378308773040771,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79442508710802,
+ "grad_norm": 0.03783095255494118,
+ "learning_rate": 6.926039770514686e-06,
+ "loss": 1.7482147216796875,
+ "step": 3513
+ },
+ {
+ "epoch": 48.80836236933798,
+ "grad_norm": 0.037651680409908295,
+ "learning_rate": 6.905123629483277e-06,
+ "loss": 1.7448663711547852,
+ "step": 3514
+ },
+ {
+ "epoch": 48.822299651567945,
+ "grad_norm": 0.03787222132086754,
+ "learning_rate": 6.884446046745608e-06,
+ "loss": 1.7379319667816162,
+ "step": 3515
+ },
+ {
+ "epoch": 48.836236933797906,
+ "grad_norm": 0.038111910223960876,
+ "learning_rate": 6.8640070389613055e-06,
+ "loss": 1.7502140998840332,
+ "step": 3516
+ },
+ {
+ "epoch": 48.850174216027874,
+ "grad_norm": 0.03940486162900925,
+ "learning_rate": 6.843806622597661e-06,
+ "loss": 1.7444273233413696,
+ "step": 3517
+ },
+ {
+ "epoch": 48.86411149825784,
+ "grad_norm": 0.03790326043963432,
+ "learning_rate": 6.823844813929796e-06,
+ "loss": 1.7383887767791748,
+ "step": 3518
+ },
+ {
+ "epoch": 48.8780487804878,
+ "grad_norm": 0.03723892569541931,
+ "learning_rate": 6.804121629040597e-06,
+ "loss": 1.740854263305664,
+ "step": 3519
+ },
+ {
+ "epoch": 48.89198606271777,
+ "grad_norm": 0.03802796080708504,
+ "learning_rate": 6.7846370838206455e-06,
+ "loss": 1.72953200340271,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90592334494774,
+ "grad_norm": 0.03777507320046425,
+ "learning_rate": 6.765391193968287e-06,
+ "loss": 1.7573529481887817,
+ "step": 3521
+ },
+ {
+ "epoch": 48.9198606271777,
+ "grad_norm": 0.03760818392038345,
+ "learning_rate": 6.7463839749896335e-06,
+ "loss": 1.744208812713623,
+ "step": 3522
+ },
+ {
+ "epoch": 48.933797909407666,
+ "grad_norm": 0.03770434111356735,
+ "learning_rate": 6.727615442198394e-06,
+ "loss": 1.7435452938079834,
+ "step": 3523
+ },
+ {
+ "epoch": 48.94773519163763,
+ "grad_norm": 0.03740346431732178,
+ "learning_rate": 6.709085610716107e-06,
+ "loss": 1.7333492040634155,
+ "step": 3524
+ },
+ {
+ "epoch": 48.961672473867594,
+ "grad_norm": 0.03788381442427635,
+ "learning_rate": 6.690794495471878e-06,
+ "loss": 1.7397178411483765,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97560975609756,
+ "grad_norm": 0.03813458979129791,
+ "learning_rate": 6.672742111202543e-06,
+ "loss": 1.7431057691574097,
+ "step": 3526
+ },
+ {
+ "epoch": 48.98954703832753,
+ "grad_norm": 0.03817913308739662,
+ "learning_rate": 6.654928472452569e-06,
+ "loss": 1.7307250499725342,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.031720418483018875,
+ "learning_rate": 6.637353593574088e-06,
+ "loss": 1.3039050102233887,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 0.660346508026123,
+ "eval_runtime": 40.794,
+ "eval_samples_per_second": 59.862,
+ "eval_steps_per_second": 0.49,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01393728222997,
+ "grad_norm": 0.03522714599967003,
+ "learning_rate": 6.620017488726867e-06,
+ "loss": 1.7501673698425293,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02787456445993,
+ "grad_norm": 0.035406239330768585,
+ "learning_rate": 6.602920171878265e-06,
+ "loss": 1.7358343601226807,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041811846689896,
+ "grad_norm": 0.03530430048704147,
+ "learning_rate": 6.5860616568033465e-06,
+ "loss": 1.7389439344406128,
+ "step": 3531
+ },
+ {
+ "epoch": 49.055749128919864,
+ "grad_norm": 0.035246510058641434,
+ "learning_rate": 6.569441957084671e-06,
+ "loss": 1.7333688735961914,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069686411149824,
+ "grad_norm": 0.03543391451239586,
+ "learning_rate": 6.5530610861124e-06,
+ "loss": 1.7434418201446533,
+ "step": 3533
+ },
+ {
+ "epoch": 49.08362369337979,
+ "grad_norm": 0.036383114755153656,
+ "learning_rate": 6.536919057084357e-06,
+ "loss": 1.7427971363067627,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09756097560975,
+ "grad_norm": 0.03543848171830177,
+ "learning_rate": 6.521015883005866e-06,
+ "loss": 1.7374334335327148,
+ "step": 3535
+ },
+ {
+ "epoch": 49.11149825783972,
+ "grad_norm": 0.03517977520823479,
+ "learning_rate": 6.5053515766898195e-06,
+ "loss": 1.7524981498718262,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12543554006969,
+ "grad_norm": 0.03517499566078186,
+ "learning_rate": 6.489926150756674e-06,
+ "loss": 1.7532435655593872,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13937282229965,
+ "grad_norm": 0.035374682396650314,
+ "learning_rate": 6.4747396176344176e-06,
+ "loss": 1.7426207065582275,
+ "step": 3538
+ },
+ {
+ "epoch": 49.153310104529616,
+ "grad_norm": 0.03599017858505249,
+ "learning_rate": 6.459791989558542e-06,
+ "loss": 1.7509880065917969,
+ "step": 3539
+ },
+ {
+ "epoch": 49.167247386759584,
+ "grad_norm": 0.0358017235994339,
+ "learning_rate": 6.445083278572134e-06,
+ "loss": 1.7249541282653809,
+ "step": 3540
+ },
+ {
+ "epoch": 49.181184668989545,
+ "grad_norm": 0.035384055227041245,
+ "learning_rate": 6.430613496525685e-06,
+ "loss": 1.7274175882339478,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19512195121951,
+ "grad_norm": 0.036022838205099106,
+ "learning_rate": 6.416382655077248e-06,
+ "loss": 1.7480113506317139,
+ "step": 3542
+ },
+ {
+ "epoch": 49.20905923344948,
+ "grad_norm": 0.03610275685787201,
+ "learning_rate": 6.402390765692381e-06,
+ "loss": 1.7519536018371582,
+ "step": 3543
+ },
+ {
+ "epoch": 49.22299651567944,
+ "grad_norm": 0.036202944815158844,
+ "learning_rate": 6.388637839644073e-06,
+ "loss": 1.747304916381836,
+ "step": 3544
+ },
+ {
+ "epoch": 49.23693379790941,
+ "grad_norm": 0.03637398034334183,
+ "learning_rate": 6.375123888012812e-06,
+ "loss": 1.7307384014129639,
+ "step": 3545
+ },
+ {
+ "epoch": 49.250871080139376,
+ "grad_norm": 0.0354069322347641,
+ "learning_rate": 6.361848921686523e-06,
+ "loss": 1.7416441440582275,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26480836236934,
+ "grad_norm": 0.03687191754579544,
+ "learning_rate": 6.348812951360662e-06,
+ "loss": 1.741818904876709,
+ "step": 3547
+ },
+ {
+ "epoch": 49.278745644599304,
+ "grad_norm": 0.03579849377274513,
+ "learning_rate": 6.33601598753802e-06,
+ "loss": 1.7351491451263428,
+ "step": 3548
+ },
+ {
+ "epoch": 49.292682926829265,
+ "grad_norm": 0.03670169785618782,
+ "learning_rate": 6.323458040528923e-06,
+ "loss": 1.736520767211914,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30662020905923,
+ "grad_norm": 0.03560818359255791,
+ "learning_rate": 6.311139120451062e-06,
+ "loss": 1.731799602508545,
+ "step": 3550
+ },
+ {
+ "epoch": 49.3205574912892,
+ "grad_norm": 0.03639044612646103,
+ "learning_rate": 6.299059237229567e-06,
+ "loss": 1.7344332933425903,
+ "step": 3551
+ },
+ {
+ "epoch": 49.33449477351916,
+ "grad_norm": 0.03560566529631615,
+ "learning_rate": 6.287218400596961e-06,
+ "loss": 1.7297918796539307,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34843205574913,
+ "grad_norm": 0.035834018141031265,
+ "learning_rate": 6.27561662009321e-06,
+ "loss": 1.742045521736145,
+ "step": 3553
+ },
+ {
+ "epoch": 49.362369337979096,
+ "grad_norm": 0.03669212758541107,
+ "learning_rate": 6.264253905065675e-06,
+ "loss": 1.7409422397613525,
+ "step": 3554
+ },
+ {
+ "epoch": 49.37630662020906,
+ "grad_norm": 0.036257825791835785,
+ "learning_rate": 6.2531302646690866e-06,
+ "loss": 1.734163761138916,
+ "step": 3555
+ },
+ {
+ "epoch": 49.390243902439025,
+ "grad_norm": 0.03602124750614166,
+ "learning_rate": 6.242245707865545e-06,
+ "loss": 1.7379150390625,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40418118466899,
+ "grad_norm": 0.037070780992507935,
+ "learning_rate": 6.23160024342455e-06,
+ "loss": 1.7512003183364868,
+ "step": 3557
+ },
+ {
+ "epoch": 49.41811846689895,
+ "grad_norm": 0.035967037081718445,
+ "learning_rate": 6.221193879922969e-06,
+ "loss": 1.7250819206237793,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43205574912892,
+ "grad_norm": 0.03709160163998604,
+ "learning_rate": 6.211026625745005e-06,
+ "loss": 1.745316743850708,
+ "step": 3559
+ },
+ {
+ "epoch": 49.44599303135889,
+ "grad_norm": 0.03603612259030342,
+ "learning_rate": 6.201098489082232e-06,
+ "loss": 1.748392939567566,
+ "step": 3560
+ },
+ {
+ "epoch": 49.45993031358885,
+ "grad_norm": 0.037297364324331284,
+ "learning_rate": 6.191409477933554e-06,
+ "loss": 1.739554762840271,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47386759581882,
+ "grad_norm": 0.03654402494430542,
+ "learning_rate": 6.18195960010528e-06,
+ "loss": 1.7408766746520996,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48780487804878,
+ "grad_norm": 0.037688035517930984,
+ "learning_rate": 6.172748863210954e-06,
+ "loss": 1.7281386852264404,
+ "step": 3563
+ },
+ {
+ "epoch": 49.501742160278745,
+ "grad_norm": 0.03664310649037361,
+ "learning_rate": 6.163777274671552e-06,
+ "loss": 1.7389543056488037,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51567944250871,
+ "grad_norm": 0.03644893318414688,
+ "learning_rate": 6.155044841715289e-06,
+ "loss": 1.7299724817276,
+ "step": 3565
+ },
+ {
+ "epoch": 49.52961672473867,
+ "grad_norm": 0.0372745618224144,
+ "learning_rate": 6.146551571377715e-06,
+ "loss": 1.7397650480270386,
+ "step": 3566
+ },
+ {
+ "epoch": 49.54355400696864,
+ "grad_norm": 0.03564862161874771,
+ "learning_rate": 6.1382974705017456e-06,
+ "loss": 1.7439970970153809,
+ "step": 3567
+ },
+ {
+ "epoch": 49.55749128919861,
+ "grad_norm": 0.03769845888018608,
+ "learning_rate": 6.130282545737535e-06,
+ "loss": 1.7407417297363281,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57142857142857,
+ "grad_norm": 0.03623263165354729,
+ "learning_rate": 6.1225068035425725e-06,
+ "loss": 1.7341506481170654,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58536585365854,
+ "grad_norm": 0.03623772785067558,
+ "learning_rate": 6.114970250181616e-06,
+ "loss": 1.7301368713378906,
+ "step": 3570
+ },
+ {
+ "epoch": 49.599303135888505,
+ "grad_norm": 0.03750680759549141,
+ "learning_rate": 6.107672891726724e-06,
+ "loss": 1.7343497276306152,
+ "step": 3571
+ },
+ {
+ "epoch": 49.613240418118465,
+ "grad_norm": 0.03659738972783089,
+ "learning_rate": 6.1006147340572916e-06,
+ "loss": 1.750082015991211,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62717770034843,
+ "grad_norm": 0.03688710927963257,
+ "learning_rate": 6.093795782859917e-06,
+ "loss": 1.735040307044983,
+ "step": 3573
+ },
+ {
+ "epoch": 49.641114982578394,
+ "grad_norm": 0.03712112829089165,
+ "learning_rate": 6.087216043628502e-06,
+ "loss": 1.7354241609573364,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65505226480836,
+ "grad_norm": 0.03706591948866844,
+ "learning_rate": 6.0808755216642146e-06,
+ "loss": 1.7423216104507446,
+ "step": 3575
+ },
+ {
+ "epoch": 49.66898954703833,
+ "grad_norm": 0.03654217720031738,
+ "learning_rate": 6.07477422207556e-06,
+ "loss": 1.7465624809265137,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68292682926829,
+ "grad_norm": 0.03709641098976135,
+ "learning_rate": 6.06891214977818e-06,
+ "loss": 1.7381062507629395,
+ "step": 3577
+ },
+ {
+ "epoch": 49.69686411149826,
+ "grad_norm": 0.03762492164969444,
+ "learning_rate": 6.063289309495051e-06,
+ "loss": 1.7440953254699707,
+ "step": 3578
+ },
+ {
+ "epoch": 49.710801393728225,
+ "grad_norm": 0.0365234836935997,
+ "learning_rate": 6.057905705756421e-06,
+ "loss": 1.7448567152023315,
+ "step": 3579
+ },
+ {
+ "epoch": 49.724738675958186,
+ "grad_norm": 0.03713703155517578,
+ "learning_rate": 6.052761342899737e-06,
+ "loss": 1.7555094957351685,
+ "step": 3580
+ },
+ {
+ "epoch": 49.73867595818815,
+ "grad_norm": 0.03783167153596878,
+ "learning_rate": 6.0478562250697185e-06,
+ "loss": 1.7434871196746826,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75261324041812,
+ "grad_norm": 0.036822400987148285,
+ "learning_rate": 6.043190356218355e-06,
+ "loss": 1.7315247058868408,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76655052264808,
+ "grad_norm": 0.03712427616119385,
+ "learning_rate": 6.038763740104835e-06,
+ "loss": 1.736377239227295,
+ "step": 3583
+ },
+ {
+ "epoch": 49.78048780487805,
+ "grad_norm": 0.03701643645763397,
+ "learning_rate": 6.034576380295586e-06,
+ "loss": 1.731011986732483,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79442508710802,
+ "grad_norm": 0.03739194571971893,
+ "learning_rate": 6.030628280164338e-06,
+ "loss": 1.7527711391448975,
+ "step": 3585
+ },
+ {
+ "epoch": 49.80836236933798,
+ "grad_norm": 0.03718961030244827,
+ "learning_rate": 6.026919442891958e-06,
+ "loss": 1.7453999519348145,
+ "step": 3586
+ },
+ {
+ "epoch": 49.822299651567945,
+ "grad_norm": 0.03757784143090248,
+ "learning_rate": 6.023449871466615e-06,
+ "loss": 1.7258204221725464,
+ "step": 3587
+ },
+ {
+ "epoch": 49.836236933797906,
+ "grad_norm": 0.0370323583483696,
+ "learning_rate": 6.020219568683678e-06,
+ "loss": 1.7448369264602661,
+ "step": 3588
+ },
+ {
+ "epoch": 49.850174216027874,
+ "grad_norm": 0.0392451211810112,
+ "learning_rate": 6.0172285371456925e-06,
+ "loss": 1.7552015781402588,
+ "step": 3589
+ },
+ {
+ "epoch": 49.86411149825784,
+ "grad_norm": 0.038492750376462936,
+ "learning_rate": 6.014476779262568e-06,
+ "loss": 1.7418248653411865,
+ "step": 3590
+ },
+ {
+ "epoch": 49.8780487804878,
+ "grad_norm": 0.03750867769122124,
+ "learning_rate": 6.011964297251285e-06,
+ "loss": 1.7437394857406616,
+ "step": 3591
+ },
+ {
+ "epoch": 49.89198606271777,
+ "grad_norm": 0.03761592134833336,
+ "learning_rate": 6.0096910931361275e-06,
+ "loss": 1.7369332313537598,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90592334494774,
+ "grad_norm": 0.03687804192304611,
+ "learning_rate": 6.007657168748551e-06,
+ "loss": 1.731837511062622,
+ "step": 3593
+ },
+ {
+ "epoch": 49.9198606271777,
+ "grad_norm": 0.037847790867090225,
+ "learning_rate": 6.005862525727277e-06,
+ "loss": 1.7474937438964844,
+ "step": 3594
+ },
+ {
+ "epoch": 49.933797909407666,
+ "grad_norm": 0.036800846457481384,
+ "learning_rate": 6.0043071655181986e-06,
+ "loss": 1.7384147644042969,
+ "step": 3595
+ },
+ {
+ "epoch": 49.94773519163763,
+ "grad_norm": 0.037281833589076996,
+ "learning_rate": 6.002991089374446e-06,
+ "loss": 1.742637038230896,
+ "step": 3596
+ },
+ {
+ "epoch": 49.961672473867594,
+ "grad_norm": 0.03809261694550514,
+ "learning_rate": 6.0019142983563875e-06,
+ "loss": 1.7374844551086426,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97560975609756,
+ "grad_norm": 0.03753192350268364,
+ "learning_rate": 6.001076793331555e-06,
+ "loss": 1.7366619110107422,
+ "step": 3598
+ },
+ {
+ "epoch": 49.98954703832753,
+ "grad_norm": 0.03671012446284294,
+ "learning_rate": 6.000478574974689e-06,
+ "loss": 1.740043044090271,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.03178739175200462,
+ "learning_rate": 6.000119643767767e-06,
+ "loss": 1.3138563632965088,
+ "step": 3600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.5388280082137088e+18,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-3600/training_args.bin b/runs/mt5-ssl/checkpoint-3600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-3600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-72/chat_template.jinja b/runs/mt5-ssl/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-72/config.json b/runs/mt5-ssl/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-72/generation_config.json b/runs/mt5-ssl/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-72/model.safetensors b/runs/mt5-ssl/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4861cd9a8c7d8679cd7c59e58db9c5c3d65d9573
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:531c46b8f5dcc4dc8d237023211981e706fc341134a2ee0d6212b856d74eb524
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-72/optimizer.pt b/runs/mt5-ssl/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..eae4346cc1db9cedc6e61f94535a624d53f42ad4
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b4af34dbd381c32cc0b19cc847b5446d4cd8321573547558f9cfc6c04ec0279a
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-72/rng_state_0.pth b/runs/mt5-ssl/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3fc68e18ddaf65dfbdec55893d9a925ff5e43a18
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4dd7671ce88d469c49c0530724ac76b2306574002d1ecd1ca9294e41621fd96a
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-72/rng_state_1.pth b/runs/mt5-ssl/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..79d9de29ae34b3b0c10ea8ba0348aeafb0c12226
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3246ef1170ccca541a03b89ad6f20e01c51eb6834a2c2211c78c71c70f896879
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-72/scheduler.pt b/runs/mt5-ssl/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b3af158e6a18004b03a150114c32a235df6d16a0
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9ad58a4676ce2fb1ab2dc9d3047a7897e6cdf70ba2536ad730ac906c3b3dd935
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-72/tokenizer.json b/runs/mt5-ssl/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-72/tokenizer_config.json b/runs/mt5-ssl/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-72/trainer_state.json b/runs/mt5-ssl/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..bac0357d679c9c510b158a15b9657281d26f208f
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.0776560164274176e+16,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-72/training_args.bin b/runs/mt5-ssl/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201
diff --git a/runs/mt5-ssl/checkpoint-720/chat_template.jinja b/runs/mt5-ssl/checkpoint-720/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/mt5-ssl/checkpoint-720/config.json b/runs/mt5-ssl/checkpoint-720/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/mt5-ssl/checkpoint-720/generation_config.json b/runs/mt5-ssl/checkpoint-720/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/mt5-ssl/checkpoint-720/model.safetensors b/runs/mt5-ssl/checkpoint-720/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a8c75ca61f711971e04850257db9717b37f582fc
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f1b5a3d20e256820c8ed62d7a9f1b5e9bac29fd149e1b289fd780ace9bad37df
+size 583356232
diff --git a/runs/mt5-ssl/checkpoint-720/optimizer.pt b/runs/mt5-ssl/checkpoint-720/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d8be2f718cb616b9ad9a6b5668c1f82eb3a8ffed
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fc221950ddce5ce095f1d8e18aa4e36dd234f7e3ee3a09cb6d6cdd418e8b0a4f
+size 1166825803
diff --git a/runs/mt5-ssl/checkpoint-720/rng_state_0.pth b/runs/mt5-ssl/checkpoint-720/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..64991ceae73e5d3389476a9b448c3cda96f73b10
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ed1b9484e9cec07c14c677aaf97ebf9dff6fa34fc12bae83c65f98c9d2c9ad0d
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-720/rng_state_1.pth b/runs/mt5-ssl/checkpoint-720/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fee3b6b64b847dcce05678311e2696b568d3185a
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d30790dc4fc4171a789605cefdf649fd85c4d4afb59efd35ee6fea409dad4915
+size 14917
diff --git a/runs/mt5-ssl/checkpoint-720/scheduler.pt b/runs/mt5-ssl/checkpoint-720/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..1ad4496e97265851e778415cb1e0b7cabc192960
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b29c543d5f10e6bbf88f194409dab3a0645a2f26aa18a377a9b03e338e95d97b
+size 1465
diff --git a/runs/mt5-ssl/checkpoint-720/tokenizer.json b/runs/mt5-ssl/checkpoint-720/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/mt5-ssl/checkpoint-720/tokenizer_config.json b/runs/mt5-ssl/checkpoint-720/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/mt5-ssl/checkpoint-720/trainer_state.json b/runs/mt5-ssl/checkpoint-720/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..d8588c139d87b07788a94d9d87ce85ca446fc4cd
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/trainer_state.json
@@ -0,0 +1,5146 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.0,
+ "eval_steps": 500,
+ "global_step": 720,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013937282229965157,
+ "grad_norm": 0.35050153732299805,
+ "learning_rate": 0.0,
+ "loss": 3.991912841796875,
+ "step": 1
+ },
+ {
+ "epoch": 0.027874564459930314,
+ "grad_norm": 0.3508828580379486,
+ "learning_rate": 5.999999999999999e-06,
+ "loss": 3.9921112060546875,
+ "step": 2
+ },
+ {
+ "epoch": 0.041811846689895474,
+ "grad_norm": 0.3439716696739197,
+ "learning_rate": 1.1999999999999999e-05,
+ "loss": 3.969581365585327,
+ "step": 3
+ },
+ {
+ "epoch": 0.05574912891986063,
+ "grad_norm": 0.31168049573898315,
+ "learning_rate": 1.7999999999999997e-05,
+ "loss": 3.927901268005371,
+ "step": 4
+ },
+ {
+ "epoch": 0.06968641114982578,
+ "grad_norm": 0.2709864675998688,
+ "learning_rate": 2.3999999999999997e-05,
+ "loss": 3.879537582397461,
+ "step": 5
+ },
+ {
+ "epoch": 0.08362369337979095,
+ "grad_norm": 0.24369819462299347,
+ "learning_rate": 2.9999999999999997e-05,
+ "loss": 3.8246335983276367,
+ "step": 6
+ },
+ {
+ "epoch": 0.0975609756097561,
+ "grad_norm": 0.197625532746315,
+ "learning_rate": 3.5999999999999994e-05,
+ "loss": 3.7721996307373047,
+ "step": 7
+ },
+ {
+ "epoch": 0.11149825783972125,
+ "grad_norm": 0.1507851928472519,
+ "learning_rate": 4.2e-05,
+ "loss": 3.730072021484375,
+ "step": 8
+ },
+ {
+ "epoch": 0.1254355400696864,
+ "grad_norm": 0.11997750401496887,
+ "learning_rate": 4.7999999999999994e-05,
+ "loss": 3.6892566680908203,
+ "step": 9
+ },
+ {
+ "epoch": 0.13937282229965156,
+ "grad_norm": 0.09249582141637802,
+ "learning_rate": 5.399999999999999e-05,
+ "loss": 3.655592441558838,
+ "step": 10
+ },
+ {
+ "epoch": 0.15331010452961671,
+ "grad_norm": 0.07137897610664368,
+ "learning_rate": 5.9999999999999995e-05,
+ "loss": 3.633018732070923,
+ "step": 11
+ },
+ {
+ "epoch": 0.1672473867595819,
+ "grad_norm": 0.0561993382871151,
+ "learning_rate": 6.599999999999999e-05,
+ "loss": 3.612427234649658,
+ "step": 12
+ },
+ {
+ "epoch": 0.18118466898954705,
+ "grad_norm": 0.04394219070672989,
+ "learning_rate": 7.199999999999999e-05,
+ "loss": 3.6006245613098145,
+ "step": 13
+ },
+ {
+ "epoch": 0.1951219512195122,
+ "grad_norm": 0.03419766575098038,
+ "learning_rate": 7.8e-05,
+ "loss": 3.5922369956970215,
+ "step": 14
+ },
+ {
+ "epoch": 0.20905923344947736,
+ "grad_norm": 0.02948337234556675,
+ "learning_rate": 8.4e-05,
+ "loss": 3.5860488414764404,
+ "step": 15
+ },
+ {
+ "epoch": 0.2229965156794425,
+ "grad_norm": 0.026799218729138374,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 3.582834243774414,
+ "step": 16
+ },
+ {
+ "epoch": 0.23693379790940766,
+ "grad_norm": 0.023768238723278046,
+ "learning_rate": 9.599999999999999e-05,
+ "loss": 3.5775821208953857,
+ "step": 17
+ },
+ {
+ "epoch": 0.2508710801393728,
+ "grad_norm": 0.02105121687054634,
+ "learning_rate": 0.000102,
+ "loss": 3.5776636600494385,
+ "step": 18
+ },
+ {
+ "epoch": 0.26480836236933797,
+ "grad_norm": 0.01891847886145115,
+ "learning_rate": 0.00010799999999999998,
+ "loss": 3.573662757873535,
+ "step": 19
+ },
+ {
+ "epoch": 0.2787456445993031,
+ "grad_norm": 0.01751410961151123,
+ "learning_rate": 0.00011399999999999999,
+ "loss": 3.5705504417419434,
+ "step": 20
+ },
+ {
+ "epoch": 0.2926829268292683,
+ "grad_norm": 0.015063339844346046,
+ "learning_rate": 0.00011999999999999999,
+ "loss": 3.5740842819213867,
+ "step": 21
+ },
+ {
+ "epoch": 0.30662020905923343,
+ "grad_norm": 0.013930639252066612,
+ "learning_rate": 0.00012599999999999997,
+ "loss": 3.567647695541382,
+ "step": 22
+ },
+ {
+ "epoch": 0.3205574912891986,
+ "grad_norm": 0.012090693227946758,
+ "learning_rate": 0.00013199999999999998,
+ "loss": 3.564181089401245,
+ "step": 23
+ },
+ {
+ "epoch": 0.3344947735191638,
+ "grad_norm": 0.011390906758606434,
+ "learning_rate": 0.000138,
+ "loss": 3.561833143234253,
+ "step": 24
+ },
+ {
+ "epoch": 0.34843205574912894,
+ "grad_norm": 0.01057762373238802,
+ "learning_rate": 0.00014399999999999998,
+ "loss": 3.5594372749328613,
+ "step": 25
+ },
+ {
+ "epoch": 0.3623693379790941,
+ "grad_norm": 0.011119714006781578,
+ "learning_rate": 0.00015,
+ "loss": 3.5524942874908447,
+ "step": 26
+ },
+ {
+ "epoch": 0.37630662020905925,
+ "grad_norm": 0.00990488100796938,
+ "learning_rate": 0.000156,
+ "loss": 3.5560035705566406,
+ "step": 27
+ },
+ {
+ "epoch": 0.3902439024390244,
+ "grad_norm": 0.010733413510024548,
+ "learning_rate": 0.000162,
+ "loss": 3.5445187091827393,
+ "step": 28
+ },
+ {
+ "epoch": 0.40418118466898956,
+ "grad_norm": 0.009585208259522915,
+ "learning_rate": 0.000168,
+ "loss": 3.550079107284546,
+ "step": 29
+ },
+ {
+ "epoch": 0.4181184668989547,
+ "grad_norm": 0.009809095412492752,
+ "learning_rate": 0.00017399999999999997,
+ "loss": 3.5437774658203125,
+ "step": 30
+ },
+ {
+ "epoch": 0.43205574912891986,
+ "grad_norm": 0.010857398621737957,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 3.54434871673584,
+ "step": 31
+ },
+ {
+ "epoch": 0.445993031358885,
+ "grad_norm": 0.00934332050383091,
+ "learning_rate": 0.000186,
+ "loss": 3.541585922241211,
+ "step": 32
+ },
+ {
+ "epoch": 0.45993031358885017,
+ "grad_norm": 0.012096591293811798,
+ "learning_rate": 0.00019199999999999998,
+ "loss": 3.5408542156219482,
+ "step": 33
+ },
+ {
+ "epoch": 0.4738675958188153,
+ "grad_norm": 0.02826732210814953,
+ "learning_rate": 0.000198,
+ "loss": 3.5377278327941895,
+ "step": 34
+ },
+ {
+ "epoch": 0.4878048780487805,
+ "grad_norm": 0.02059389464557171,
+ "learning_rate": 0.000204,
+ "loss": 3.5415492057800293,
+ "step": 35
+ },
+ {
+ "epoch": 0.5017421602787456,
+ "grad_norm": 0.021392373368144035,
+ "learning_rate": 0.00020999999999999998,
+ "loss": 3.531378746032715,
+ "step": 36
+ },
+ {
+ "epoch": 0.5156794425087108,
+ "grad_norm": 0.016397396102547646,
+ "learning_rate": 0.00021599999999999996,
+ "loss": 3.5238890647888184,
+ "step": 37
+ },
+ {
+ "epoch": 0.5296167247386759,
+ "grad_norm": 0.022605035454034805,
+ "learning_rate": 0.00022199999999999998,
+ "loss": 3.5180482864379883,
+ "step": 38
+ },
+ {
+ "epoch": 0.5435540069686411,
+ "grad_norm": 0.027316192165017128,
+ "learning_rate": 0.00022799999999999999,
+ "loss": 3.5050134658813477,
+ "step": 39
+ },
+ {
+ "epoch": 0.5574912891986062,
+ "grad_norm": 0.020708877593278885,
+ "learning_rate": 0.000234,
+ "loss": 3.499526023864746,
+ "step": 40
+ },
+ {
+ "epoch": 0.5714285714285714,
+ "grad_norm": 0.03667809069156647,
+ "learning_rate": 0.00023999999999999998,
+ "loss": 3.4915785789489746,
+ "step": 41
+ },
+ {
+ "epoch": 0.5853658536585366,
+ "grad_norm": 0.02999032475054264,
+ "learning_rate": 0.00024599999999999996,
+ "loss": 3.480621099472046,
+ "step": 42
+ },
+ {
+ "epoch": 0.5993031358885017,
+ "grad_norm": 0.028212690725922585,
+ "learning_rate": 0.00025199999999999995,
+ "loss": 3.4665350914001465,
+ "step": 43
+ },
+ {
+ "epoch": 0.6132404181184669,
+ "grad_norm": 0.022621743381023407,
+ "learning_rate": 0.000258,
+ "loss": 3.4697375297546387,
+ "step": 44
+ },
+ {
+ "epoch": 0.627177700348432,
+ "grad_norm": 0.03026035614311695,
+ "learning_rate": 0.00026399999999999997,
+ "loss": 3.4668068885803223,
+ "step": 45
+ },
+ {
+ "epoch": 0.6411149825783972,
+ "grad_norm": 0.03631209582090378,
+ "learning_rate": 0.00027,
+ "loss": 3.4558091163635254,
+ "step": 46
+ },
+ {
+ "epoch": 0.6550522648083623,
+ "grad_norm": 0.03253362327814102,
+ "learning_rate": 0.000276,
+ "loss": 3.4463515281677246,
+ "step": 47
+ },
+ {
+ "epoch": 0.6689895470383276,
+ "grad_norm": 0.028335504233837128,
+ "learning_rate": 0.00028199999999999997,
+ "loss": 3.437135934829712,
+ "step": 48
+ },
+ {
+ "epoch": 0.6829268292682927,
+ "grad_norm": 0.06646716594696045,
+ "learning_rate": 0.00028799999999999995,
+ "loss": 3.428346872329712,
+ "step": 49
+ },
+ {
+ "epoch": 0.6968641114982579,
+ "grad_norm": 0.07188845425844193,
+ "learning_rate": 0.000294,
+ "loss": 3.430262327194214,
+ "step": 50
+ },
+ {
+ "epoch": 0.710801393728223,
+ "grad_norm": 0.03682823106646538,
+ "learning_rate": 0.0003,
+ "loss": 3.416962146759033,
+ "step": 51
+ },
+ {
+ "epoch": 0.7247386759581882,
+ "grad_norm": 0.0755256861448288,
+ "learning_rate": 0.00030599999999999996,
+ "loss": 3.4191060066223145,
+ "step": 52
+ },
+ {
+ "epoch": 0.7386759581881533,
+ "grad_norm": 0.1758236289024353,
+ "learning_rate": 0.000312,
+ "loss": 3.4200382232666016,
+ "step": 53
+ },
+ {
+ "epoch": 0.7526132404181185,
+ "grad_norm": 0.15794925391674042,
+ "learning_rate": 0.000318,
+ "loss": 3.4228811264038086,
+ "step": 54
+ },
+ {
+ "epoch": 0.7665505226480837,
+ "grad_norm": 0.026228366419672966,
+ "learning_rate": 0.000324,
+ "loss": 3.3970746994018555,
+ "step": 55
+ },
+ {
+ "epoch": 0.7804878048780488,
+ "grad_norm": 0.08065437525510788,
+ "learning_rate": 0.00033,
+ "loss": 3.402804136276245,
+ "step": 56
+ },
+ {
+ "epoch": 0.794425087108014,
+ "grad_norm": 0.018620377406477928,
+ "learning_rate": 0.000336,
+ "loss": 3.3972244262695312,
+ "step": 57
+ },
+ {
+ "epoch": 0.8083623693379791,
+ "grad_norm": 0.06154120713472366,
+ "learning_rate": 0.00034199999999999996,
+ "loss": 3.3930704593658447,
+ "step": 58
+ },
+ {
+ "epoch": 0.8222996515679443,
+ "grad_norm": 0.0240298081189394,
+ "learning_rate": 0.00034799999999999995,
+ "loss": 3.3866734504699707,
+ "step": 59
+ },
+ {
+ "epoch": 0.8362369337979094,
+ "grad_norm": 0.05287991091609001,
+ "learning_rate": 0.00035399999999999993,
+ "loss": 3.3821773529052734,
+ "step": 60
+ },
+ {
+ "epoch": 0.8501742160278746,
+ "grad_norm": 0.037251003086566925,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 3.3710193634033203,
+ "step": 61
+ },
+ {
+ "epoch": 0.8641114982578397,
+ "grad_norm": 0.05490389093756676,
+ "learning_rate": 0.00036599999999999995,
+ "loss": 3.3778605461120605,
+ "step": 62
+ },
+ {
+ "epoch": 0.8780487804878049,
+ "grad_norm": 0.03813353553414345,
+ "learning_rate": 0.000372,
+ "loss": 3.3674263954162598,
+ "step": 63
+ },
+ {
+ "epoch": 0.89198606271777,
+ "grad_norm": 0.05060660466551781,
+ "learning_rate": 0.00037799999999999997,
+ "loss": 3.3638195991516113,
+ "step": 64
+ },
+ {
+ "epoch": 0.9059233449477352,
+ "grad_norm": 0.044462986290454865,
+ "learning_rate": 0.00038399999999999996,
+ "loss": 3.3628158569335938,
+ "step": 65
+ },
+ {
+ "epoch": 0.9198606271777003,
+ "grad_norm": 0.034862492233514786,
+ "learning_rate": 0.00039,
+ "loss": 3.359663963317871,
+ "step": 66
+ },
+ {
+ "epoch": 0.9337979094076655,
+ "grad_norm": 0.04654683545231819,
+ "learning_rate": 0.000396,
+ "loss": 3.355091094970703,
+ "step": 67
+ },
+ {
+ "epoch": 0.9477351916376306,
+ "grad_norm": 0.041474517434835434,
+ "learning_rate": 0.000402,
+ "loss": 3.3407883644104004,
+ "step": 68
+ },
+ {
+ "epoch": 0.9616724738675958,
+ "grad_norm": 0.043404098600149155,
+ "learning_rate": 0.000408,
+ "loss": 3.3388497829437256,
+ "step": 69
+ },
+ {
+ "epoch": 0.975609756097561,
+ "grad_norm": 0.050107456743717194,
+ "learning_rate": 0.0004139999999999999,
+ "loss": 3.326899766921997,
+ "step": 70
+ },
+ {
+ "epoch": 0.9895470383275261,
+ "grad_norm": 0.022057073190808296,
+ "learning_rate": 0.00041999999999999996,
+ "loss": 3.321089744567871,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.03693892061710358,
+ "learning_rate": 0.00042599999999999995,
+ "loss": 2.491826295852661,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 0.83028644323349,
+ "eval_runtime": 40.3196,
+ "eval_samples_per_second": 60.566,
+ "eval_steps_per_second": 0.496,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139372822299653,
+ "grad_norm": 0.03299665451049805,
+ "learning_rate": 0.00043199999999999993,
+ "loss": 3.312657117843628,
+ "step": 73
+ },
+ {
+ "epoch": 1.0278745644599303,
+ "grad_norm": 0.05104782059788704,
+ "learning_rate": 0.00043799999999999997,
+ "loss": 3.3143444061279297,
+ "step": 74
+ },
+ {
+ "epoch": 1.0418118466898956,
+ "grad_norm": 0.050383053719997406,
+ "learning_rate": 0.00044399999999999995,
+ "loss": 3.3019282817840576,
+ "step": 75
+ },
+ {
+ "epoch": 1.0557491289198606,
+ "grad_norm": 0.08193834125995636,
+ "learning_rate": 0.00045,
+ "loss": 3.3052425384521484,
+ "step": 76
+ },
+ {
+ "epoch": 1.0696864111498259,
+ "grad_norm": 0.13088618218898773,
+ "learning_rate": 0.00045599999999999997,
+ "loss": 3.304067611694336,
+ "step": 77
+ },
+ {
+ "epoch": 1.083623693379791,
+ "grad_norm": 0.14614517986774445,
+ "learning_rate": 0.00046199999999999995,
+ "loss": 3.298491954803467,
+ "step": 78
+ },
+ {
+ "epoch": 1.0975609756097562,
+ "grad_norm": 0.2019524723291397,
+ "learning_rate": 0.000468,
+ "loss": 3.3283910751342773,
+ "step": 79
+ },
+ {
+ "epoch": 1.1114982578397212,
+ "grad_norm": 0.10340095311403275,
+ "learning_rate": 0.000474,
+ "loss": 3.3009772300720215,
+ "step": 80
+ },
+ {
+ "epoch": 1.1254355400696865,
+ "grad_norm": 0.09748320281505585,
+ "learning_rate": 0.00047999999999999996,
+ "loss": 3.2933666706085205,
+ "step": 81
+ },
+ {
+ "epoch": 1.1393728222996515,
+ "grad_norm": 0.08568419516086578,
+ "learning_rate": 0.000486,
+ "loss": 3.2799019813537598,
+ "step": 82
+ },
+ {
+ "epoch": 1.1533101045296168,
+ "grad_norm": 0.07254650443792343,
+ "learning_rate": 0.0004919999999999999,
+ "loss": 3.2744603157043457,
+ "step": 83
+ },
+ {
+ "epoch": 1.1672473867595818,
+ "grad_norm": 0.09184414148330688,
+ "learning_rate": 0.000498,
+ "loss": 3.265300750732422,
+ "step": 84
+ },
+ {
+ "epoch": 1.181184668989547,
+ "grad_norm": 0.057995934039354324,
+ "learning_rate": 0.0005039999999999999,
+ "loss": 3.260770797729492,
+ "step": 85
+ },
+ {
+ "epoch": 1.1951219512195121,
+ "grad_norm": 0.07937615364789963,
+ "learning_rate": 0.0005099999999999999,
+ "loss": 3.257512092590332,
+ "step": 86
+ },
+ {
+ "epoch": 1.2090592334494774,
+ "grad_norm": 0.05541674420237541,
+ "learning_rate": 0.000516,
+ "loss": 3.2583250999450684,
+ "step": 87
+ },
+ {
+ "epoch": 1.2229965156794425,
+ "grad_norm": 0.053845975548028946,
+ "learning_rate": 0.000522,
+ "loss": 3.2531747817993164,
+ "step": 88
+ },
+ {
+ "epoch": 1.2369337979094077,
+ "grad_norm": 0.08838699012994766,
+ "learning_rate": 0.0005279999999999999,
+ "loss": 3.241713523864746,
+ "step": 89
+ },
+ {
+ "epoch": 1.2508710801393728,
+ "grad_norm": 0.10236947983503342,
+ "learning_rate": 0.000534,
+ "loss": 3.2416672706604004,
+ "step": 90
+ },
+ {
+ "epoch": 1.264808362369338,
+ "grad_norm": 0.10996634513139725,
+ "learning_rate": 0.00054,
+ "loss": 3.244565963745117,
+ "step": 91
+ },
+ {
+ "epoch": 1.278745644599303,
+ "grad_norm": 0.0631512850522995,
+ "learning_rate": 0.0005459999999999999,
+ "loss": 3.2318668365478516,
+ "step": 92
+ },
+ {
+ "epoch": 1.2926829268292683,
+ "grad_norm": 0.0588001124560833,
+ "learning_rate": 0.000552,
+ "loss": 3.228060245513916,
+ "step": 93
+ },
+ {
+ "epoch": 1.3066202090592334,
+ "grad_norm": 0.06863059103488922,
+ "learning_rate": 0.000558,
+ "loss": 3.215839385986328,
+ "step": 94
+ },
+ {
+ "epoch": 1.3205574912891986,
+ "grad_norm": 0.03364138305187225,
+ "learning_rate": 0.0005639999999999999,
+ "loss": 3.2204508781433105,
+ "step": 95
+ },
+ {
+ "epoch": 1.3344947735191637,
+ "grad_norm": 0.0730782151222229,
+ "learning_rate": 0.00057,
+ "loss": 3.2238316535949707,
+ "step": 96
+ },
+ {
+ "epoch": 1.348432055749129,
+ "grad_norm": 0.055765505880117416,
+ "learning_rate": 0.0005759999999999999,
+ "loss": 3.2094883918762207,
+ "step": 97
+ },
+ {
+ "epoch": 1.3623693379790942,
+ "grad_norm": 0.04374352842569351,
+ "learning_rate": 0.0005819999999999999,
+ "loss": 3.2067670822143555,
+ "step": 98
+ },
+ {
+ "epoch": 1.3763066202090593,
+ "grad_norm": 0.052591972053050995,
+ "learning_rate": 0.000588,
+ "loss": 3.201781749725342,
+ "step": 99
+ },
+ {
+ "epoch": 1.3902439024390243,
+ "grad_norm": 0.043141428381204605,
+ "learning_rate": 0.0005939999999999999,
+ "loss": 3.192054271697998,
+ "step": 100
+ },
+ {
+ "epoch": 1.4041811846689896,
+ "grad_norm": 0.0641680657863617,
+ "learning_rate": 0.0006,
+ "loss": 3.198087692260742,
+ "step": 101
+ },
+ {
+ "epoch": 1.4181184668989548,
+ "grad_norm": 0.04505930840969086,
+ "learning_rate": 0.0005999998803562321,
+ "loss": 3.1883275508880615,
+ "step": 102
+ },
+ {
+ "epoch": 1.4320557491289199,
+ "grad_norm": 0.04068157449364662,
+ "learning_rate": 0.0005999995214250253,
+ "loss": 3.1804330348968506,
+ "step": 103
+ },
+ {
+ "epoch": 1.445993031358885,
+ "grad_norm": 0.03246529772877693,
+ "learning_rate": 0.0005999989232066684,
+ "loss": 3.1744279861450195,
+ "step": 104
+ },
+ {
+ "epoch": 1.4599303135888502,
+ "grad_norm": 0.04536713287234306,
+ "learning_rate": 0.0005999980857016436,
+ "loss": 3.1696314811706543,
+ "step": 105
+ },
+ {
+ "epoch": 1.4738675958188154,
+ "grad_norm": 0.05856936424970627,
+ "learning_rate": 0.0005999970089106256,
+ "loss": 3.1651716232299805,
+ "step": 106
+ },
+ {
+ "epoch": 1.4878048780487805,
+ "grad_norm": 0.04254481941461563,
+ "learning_rate": 0.0005999956928344818,
+ "loss": 3.162069320678711,
+ "step": 107
+ },
+ {
+ "epoch": 1.5017421602787455,
+ "grad_norm": 0.06294365227222443,
+ "learning_rate": 0.0005999941374742726,
+ "loss": 3.1580231189727783,
+ "step": 108
+ },
+ {
+ "epoch": 1.5156794425087108,
+ "grad_norm": 0.06742165237665176,
+ "learning_rate": 0.0005999923428312514,
+ "loss": 3.1509876251220703,
+ "step": 109
+ },
+ {
+ "epoch": 1.529616724738676,
+ "grad_norm": 0.07527563720941544,
+ "learning_rate": 0.0005999903089068637,
+ "loss": 3.1563079357147217,
+ "step": 110
+ },
+ {
+ "epoch": 1.543554006968641,
+ "grad_norm": 0.09003929048776627,
+ "learning_rate": 0.0005999880357027487,
+ "loss": 3.1501898765563965,
+ "step": 111
+ },
+ {
+ "epoch": 1.5574912891986061,
+ "grad_norm": 0.0827120840549469,
+ "learning_rate": 0.0005999855232207374,
+ "loss": 3.1430892944335938,
+ "step": 112
+ },
+ {
+ "epoch": 1.5714285714285714,
+ "grad_norm": 0.06946796923875809,
+ "learning_rate": 0.0005999827714628542,
+ "loss": 3.1321616172790527,
+ "step": 113
+ },
+ {
+ "epoch": 1.5853658536585367,
+ "grad_norm": 0.10428689420223236,
+ "learning_rate": 0.0005999797804313163,
+ "loss": 3.1389565467834473,
+ "step": 114
+ },
+ {
+ "epoch": 1.5993031358885017,
+ "grad_norm": 0.11353038996458054,
+ "learning_rate": 0.0005999765501285333,
+ "loss": 3.1335158348083496,
+ "step": 115
+ },
+ {
+ "epoch": 1.6132404181184667,
+ "grad_norm": 0.12943491339683533,
+ "learning_rate": 0.000599973080557108,
+ "loss": 3.146712064743042,
+ "step": 116
+ },
+ {
+ "epoch": 1.627177700348432,
+ "grad_norm": 0.1294851303100586,
+ "learning_rate": 0.0005999693717198356,
+ "loss": 3.1308398246765137,
+ "step": 117
+ },
+ {
+ "epoch": 1.6411149825783973,
+ "grad_norm": 0.06117495149374008,
+ "learning_rate": 0.0005999654236197044,
+ "loss": 3.115572929382324,
+ "step": 118
+ },
+ {
+ "epoch": 1.6550522648083623,
+ "grad_norm": 0.08896767348051071,
+ "learning_rate": 0.0005999612362598951,
+ "loss": 3.1199288368225098,
+ "step": 119
+ },
+ {
+ "epoch": 1.6689895470383276,
+ "grad_norm": 0.07132359594106674,
+ "learning_rate": 0.0005999568096437816,
+ "loss": 3.1066761016845703,
+ "step": 120
+ },
+ {
+ "epoch": 1.6829268292682928,
+ "grad_norm": 0.05068275332450867,
+ "learning_rate": 0.0005999521437749303,
+ "loss": 3.100792407989502,
+ "step": 121
+ },
+ {
+ "epoch": 1.6968641114982579,
+ "grad_norm": 0.09045132994651794,
+ "learning_rate": 0.0005999472386571002,
+ "loss": 3.112086296081543,
+ "step": 122
+ },
+ {
+ "epoch": 1.710801393728223,
+ "grad_norm": 0.07954756915569305,
+ "learning_rate": 0.0005999420942942435,
+ "loss": 3.1000494956970215,
+ "step": 123
+ },
+ {
+ "epoch": 1.7247386759581882,
+ "grad_norm": 0.046099890023469925,
+ "learning_rate": 0.000599936710690505,
+ "loss": 3.1002321243286133,
+ "step": 124
+ },
+ {
+ "epoch": 1.7386759581881535,
+ "grad_norm": 0.057978615164756775,
+ "learning_rate": 0.0005999310878502218,
+ "loss": 3.0907797813415527,
+ "step": 125
+ },
+ {
+ "epoch": 1.7526132404181185,
+ "grad_norm": 0.05730621889233589,
+ "learning_rate": 0.0005999252257779244,
+ "loss": 3.0988454818725586,
+ "step": 126
+ },
+ {
+ "epoch": 1.7665505226480835,
+ "grad_norm": 0.03337423503398895,
+ "learning_rate": 0.0005999191244783357,
+ "loss": 3.0923471450805664,
+ "step": 127
+ },
+ {
+ "epoch": 1.7804878048780488,
+ "grad_norm": 0.05005614832043648,
+ "learning_rate": 0.0005999127839563715,
+ "loss": 3.0876643657684326,
+ "step": 128
+ },
+ {
+ "epoch": 1.794425087108014,
+ "grad_norm": 0.04342116415500641,
+ "learning_rate": 0.00059990620421714,
+ "loss": 3.0790348052978516,
+ "step": 129
+ },
+ {
+ "epoch": 1.8083623693379791,
+ "grad_norm": 0.07091791927814484,
+ "learning_rate": 0.0005998993852659426,
+ "loss": 3.082573413848877,
+ "step": 130
+ },
+ {
+ "epoch": 1.8222996515679442,
+ "grad_norm": 0.07982946932315826,
+ "learning_rate": 0.0005998923271082733,
+ "loss": 3.083280086517334,
+ "step": 131
+ },
+ {
+ "epoch": 1.8362369337979094,
+ "grad_norm": 0.060250554233789444,
+ "learning_rate": 0.0005998850297498183,
+ "loss": 3.0605459213256836,
+ "step": 132
+ },
+ {
+ "epoch": 1.8501742160278747,
+ "grad_norm": 0.05904680863022804,
+ "learning_rate": 0.0005998774931964574,
+ "loss": 3.0623326301574707,
+ "step": 133
+ },
+ {
+ "epoch": 1.8641114982578397,
+ "grad_norm": 0.04342680424451828,
+ "learning_rate": 0.0005998697174542624,
+ "loss": 3.0587317943573,
+ "step": 134
+ },
+ {
+ "epoch": 1.8780487804878048,
+ "grad_norm": 0.04994596168398857,
+ "learning_rate": 0.0005998617025294983,
+ "loss": 3.054107904434204,
+ "step": 135
+ },
+ {
+ "epoch": 1.89198606271777,
+ "grad_norm": 0.0626574233174324,
+ "learning_rate": 0.0005998534484286223,
+ "loss": 3.055121421813965,
+ "step": 136
+ },
+ {
+ "epoch": 1.9059233449477353,
+ "grad_norm": 0.07644608616828918,
+ "learning_rate": 0.0005998449551582847,
+ "loss": 3.0568695068359375,
+ "step": 137
+ },
+ {
+ "epoch": 1.9198606271777003,
+ "grad_norm": 0.09572888165712357,
+ "learning_rate": 0.0005998362227253285,
+ "loss": 3.0544114112854004,
+ "step": 138
+ },
+ {
+ "epoch": 1.9337979094076654,
+ "grad_norm": 0.11117683351039886,
+ "learning_rate": 0.0005998272511367889,
+ "loss": 3.0536742210388184,
+ "step": 139
+ },
+ {
+ "epoch": 1.9477351916376306,
+ "grad_norm": 0.10240641236305237,
+ "learning_rate": 0.0005998180403998947,
+ "loss": 3.0611352920532227,
+ "step": 140
+ },
+ {
+ "epoch": 1.961672473867596,
+ "grad_norm": 0.04600656405091286,
+ "learning_rate": 0.0005998085905220664,
+ "loss": 3.0386037826538086,
+ "step": 141
+ },
+ {
+ "epoch": 1.975609756097561,
+ "grad_norm": 0.08889233320951462,
+ "learning_rate": 0.0005997989015109177,
+ "loss": 3.039328098297119,
+ "step": 142
+ },
+ {
+ "epoch": 1.989547038327526,
+ "grad_norm": 0.06673125922679901,
+ "learning_rate": 0.000599788973374255,
+ "loss": 3.03155255317688,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.038104113191366196,
+ "learning_rate": 0.000599778806120077,
+ "loss": 2.2694902420043945,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 0.7617935538291931,
+ "eval_runtime": 40.5787,
+ "eval_samples_per_second": 60.179,
+ "eval_steps_per_second": 0.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.013937282229965,
+ "grad_norm": 0.05889542028307915,
+ "learning_rate": 0.0005997683997565754,
+ "loss": 3.027337074279785,
+ "step": 145
+ },
+ {
+ "epoch": 2.0278745644599305,
+ "grad_norm": 0.07518725097179413,
+ "learning_rate": 0.0005997577542921344,
+ "loss": 3.0170812606811523,
+ "step": 146
+ },
+ {
+ "epoch": 2.0418118466898956,
+ "grad_norm": 0.0896688774228096,
+ "learning_rate": 0.0005997468697353309,
+ "loss": 3.0275402069091797,
+ "step": 147
+ },
+ {
+ "epoch": 2.0557491289198606,
+ "grad_norm": 0.051436759531497955,
+ "learning_rate": 0.0005997357460949342,
+ "loss": 3.0196452140808105,
+ "step": 148
+ },
+ {
+ "epoch": 2.0696864111498257,
+ "grad_norm": 0.050903454422950745,
+ "learning_rate": 0.0005997243833799068,
+ "loss": 3.0259199142456055,
+ "step": 149
+ },
+ {
+ "epoch": 2.083623693379791,
+ "grad_norm": 0.04333338141441345,
+ "learning_rate": 0.000599712781599403,
+ "loss": 3.017059803009033,
+ "step": 150
+ },
+ {
+ "epoch": 2.097560975609756,
+ "grad_norm": 0.04208024963736534,
+ "learning_rate": 0.0005997009407627704,
+ "loss": 3.015444755554199,
+ "step": 151
+ },
+ {
+ "epoch": 2.1114982578397212,
+ "grad_norm": 0.037582479417324066,
+ "learning_rate": 0.000599688860879549,
+ "loss": 3.0063562393188477,
+ "step": 152
+ },
+ {
+ "epoch": 2.1254355400696863,
+ "grad_norm": 0.034189529716968536,
+ "learning_rate": 0.0005996765419594711,
+ "loss": 3.001277446746826,
+ "step": 153
+ },
+ {
+ "epoch": 2.1393728222996518,
+ "grad_norm": 0.03017655573785305,
+ "learning_rate": 0.000599663984012462,
+ "loss": 2.9961438179016113,
+ "step": 154
+ },
+ {
+ "epoch": 2.153310104529617,
+ "grad_norm": 0.03874786198139191,
+ "learning_rate": 0.0005996511870486393,
+ "loss": 3.0001797676086426,
+ "step": 155
+ },
+ {
+ "epoch": 2.167247386759582,
+ "grad_norm": 0.060266636312007904,
+ "learning_rate": 0.0005996381510783135,
+ "loss": 2.993195056915283,
+ "step": 156
+ },
+ {
+ "epoch": 2.181184668989547,
+ "grad_norm": 0.10270293802022934,
+ "learning_rate": 0.0005996248761119872,
+ "loss": 3.0006868839263916,
+ "step": 157
+ },
+ {
+ "epoch": 2.1951219512195124,
+ "grad_norm": 0.1447659581899643,
+ "learning_rate": 0.0005996113621603559,
+ "loss": 3.0025670528411865,
+ "step": 158
+ },
+ {
+ "epoch": 2.2090592334494774,
+ "grad_norm": 0.12111099809408188,
+ "learning_rate": 0.0005995976092343075,
+ "loss": 3.0041494369506836,
+ "step": 159
+ },
+ {
+ "epoch": 2.2229965156794425,
+ "grad_norm": 0.06567822396755219,
+ "learning_rate": 0.0005995836173449227,
+ "loss": 2.998570442199707,
+ "step": 160
+ },
+ {
+ "epoch": 2.2369337979094075,
+ "grad_norm": 0.061619363725185394,
+ "learning_rate": 0.0005995693865034743,
+ "loss": 2.9947350025177,
+ "step": 161
+ },
+ {
+ "epoch": 2.250871080139373,
+ "grad_norm": 0.06767922639846802,
+ "learning_rate": 0.0005995549167214278,
+ "loss": 2.9967260360717773,
+ "step": 162
+ },
+ {
+ "epoch": 2.264808362369338,
+ "grad_norm": 0.062205322086811066,
+ "learning_rate": 0.0005995402080104414,
+ "loss": 2.993173122406006,
+ "step": 163
+ },
+ {
+ "epoch": 2.278745644599303,
+ "grad_norm": 0.0801854133605957,
+ "learning_rate": 0.0005995252603823656,
+ "loss": 2.99283504486084,
+ "step": 164
+ },
+ {
+ "epoch": 2.292682926829268,
+ "grad_norm": 0.11149590462446213,
+ "learning_rate": 0.0005995100738492433,
+ "loss": 2.992608070373535,
+ "step": 165
+ },
+ {
+ "epoch": 2.3066202090592336,
+ "grad_norm": 0.09314344823360443,
+ "learning_rate": 0.0005994946484233102,
+ "loss": 2.9852240085601807,
+ "step": 166
+ },
+ {
+ "epoch": 2.3205574912891986,
+ "grad_norm": 0.04199904948472977,
+ "learning_rate": 0.0005994789841169941,
+ "loss": 2.98305606842041,
+ "step": 167
+ },
+ {
+ "epoch": 2.3344947735191637,
+ "grad_norm": 0.04643206670880318,
+ "learning_rate": 0.0005994630809429156,
+ "loss": 2.9819021224975586,
+ "step": 168
+ },
+ {
+ "epoch": 2.3484320557491287,
+ "grad_norm": 0.03310038521885872,
+ "learning_rate": 0.0005994469389138875,
+ "loss": 2.975208282470703,
+ "step": 169
+ },
+ {
+ "epoch": 2.362369337979094,
+ "grad_norm": 0.053549010306596756,
+ "learning_rate": 0.0005994305580429152,
+ "loss": 2.9745559692382812,
+ "step": 170
+ },
+ {
+ "epoch": 2.3763066202090593,
+ "grad_norm": 0.03352247178554535,
+ "learning_rate": 0.0005994139383431966,
+ "loss": 2.967108726501465,
+ "step": 171
+ },
+ {
+ "epoch": 2.3902439024390243,
+ "grad_norm": 0.04210606589913368,
+ "learning_rate": 0.0005993970798281218,
+ "loss": 2.959658622741699,
+ "step": 172
+ },
+ {
+ "epoch": 2.40418118466899,
+ "grad_norm": 0.03518713638186455,
+ "learning_rate": 0.000599379982511273,
+ "loss": 2.966585874557495,
+ "step": 173
+ },
+ {
+ "epoch": 2.418118466898955,
+ "grad_norm": 0.043215107172727585,
+ "learning_rate": 0.0005993626464064259,
+ "loss": 2.9637069702148438,
+ "step": 174
+ },
+ {
+ "epoch": 2.43205574912892,
+ "grad_norm": 0.034570302814245224,
+ "learning_rate": 0.0005993450715275474,
+ "loss": 2.9604549407958984,
+ "step": 175
+ },
+ {
+ "epoch": 2.445993031358885,
+ "grad_norm": 0.031179504469037056,
+ "learning_rate": 0.0005993272578887974,
+ "loss": 2.963344097137451,
+ "step": 176
+ },
+ {
+ "epoch": 2.45993031358885,
+ "grad_norm": 0.03287631273269653,
+ "learning_rate": 0.000599309205504528,
+ "loss": 2.956332206726074,
+ "step": 177
+ },
+ {
+ "epoch": 2.4738675958188154,
+ "grad_norm": 0.036427177488803864,
+ "learning_rate": 0.0005992909143892837,
+ "loss": 2.95066237449646,
+ "step": 178
+ },
+ {
+ "epoch": 2.4878048780487805,
+ "grad_norm": 0.038861170411109924,
+ "learning_rate": 0.0005992723845578016,
+ "loss": 2.948594570159912,
+ "step": 179
+ },
+ {
+ "epoch": 2.5017421602787455,
+ "grad_norm": 0.04600491374731064,
+ "learning_rate": 0.0005992536160250103,
+ "loss": 2.9483413696289062,
+ "step": 180
+ },
+ {
+ "epoch": 2.515679442508711,
+ "grad_norm": 0.06583918631076813,
+ "learning_rate": 0.0005992346088060317,
+ "loss": 2.9452760219573975,
+ "step": 181
+ },
+ {
+ "epoch": 2.529616724738676,
+ "grad_norm": 0.07678504288196564,
+ "learning_rate": 0.0005992153629161793,
+ "loss": 2.950690269470215,
+ "step": 182
+ },
+ {
+ "epoch": 2.543554006968641,
+ "grad_norm": 0.07686694711446762,
+ "learning_rate": 0.0005991958783709593,
+ "loss": 2.957883358001709,
+ "step": 183
+ },
+ {
+ "epoch": 2.557491289198606,
+ "grad_norm": 0.057035740464925766,
+ "learning_rate": 0.0005991761551860702,
+ "loss": 2.941667318344116,
+ "step": 184
+ },
+ {
+ "epoch": 2.571428571428571,
+ "grad_norm": 0.03659744933247566,
+ "learning_rate": 0.0005991561933774022,
+ "loss": 2.9398202896118164,
+ "step": 185
+ },
+ {
+ "epoch": 2.5853658536585367,
+ "grad_norm": 0.05060099810361862,
+ "learning_rate": 0.0005991359929610386,
+ "loss": 2.9242119789123535,
+ "step": 186
+ },
+ {
+ "epoch": 2.5993031358885017,
+ "grad_norm": 0.05470975860953331,
+ "learning_rate": 0.0005991155539532544,
+ "loss": 2.933957815170288,
+ "step": 187
+ },
+ {
+ "epoch": 2.6132404181184667,
+ "grad_norm": 0.043858081102371216,
+ "learning_rate": 0.0005990948763705167,
+ "loss": 2.93229341506958,
+ "step": 188
+ },
+ {
+ "epoch": 2.6271777003484322,
+ "grad_norm": 0.04652675986289978,
+ "learning_rate": 0.0005990739602294852,
+ "loss": 2.9253902435302734,
+ "step": 189
+ },
+ {
+ "epoch": 2.6411149825783973,
+ "grad_norm": 0.06412844359874725,
+ "learning_rate": 0.0005990528055470118,
+ "loss": 2.9311153888702393,
+ "step": 190
+ },
+ {
+ "epoch": 2.6550522648083623,
+ "grad_norm": 0.08587725460529327,
+ "learning_rate": 0.0005990314123401403,
+ "loss": 2.924844264984131,
+ "step": 191
+ },
+ {
+ "epoch": 2.6689895470383274,
+ "grad_norm": 0.10669463127851486,
+ "learning_rate": 0.0005990097806261068,
+ "loss": 2.93114972114563,
+ "step": 192
+ },
+ {
+ "epoch": 2.682926829268293,
+ "grad_norm": 0.07489299774169922,
+ "learning_rate": 0.0005989879104223397,
+ "loss": 2.9279603958129883,
+ "step": 193
+ },
+ {
+ "epoch": 2.696864111498258,
+ "grad_norm": 0.07560845464468002,
+ "learning_rate": 0.0005989658017464593,
+ "loss": 2.9288363456726074,
+ "step": 194
+ },
+ {
+ "epoch": 2.710801393728223,
+ "grad_norm": 0.085471510887146,
+ "learning_rate": 0.0005989434546162782,
+ "loss": 2.932122230529785,
+ "step": 195
+ },
+ {
+ "epoch": 2.7247386759581884,
+ "grad_norm": 0.09120018035173416,
+ "learning_rate": 0.0005989208690498012,
+ "loss": 2.932145595550537,
+ "step": 196
+ },
+ {
+ "epoch": 2.7386759581881535,
+ "grad_norm": 0.06665205955505371,
+ "learning_rate": 0.0005988980450652248,
+ "loss": 2.9127566814422607,
+ "step": 197
+ },
+ {
+ "epoch": 2.7526132404181185,
+ "grad_norm": 0.056125201284885406,
+ "learning_rate": 0.0005988749826809381,
+ "loss": 2.898498058319092,
+ "step": 198
+ },
+ {
+ "epoch": 2.7665505226480835,
+ "grad_norm": 0.046871136873960495,
+ "learning_rate": 0.0005988516819155222,
+ "loss": 2.913458824157715,
+ "step": 199
+ },
+ {
+ "epoch": 2.7804878048780486,
+ "grad_norm": 0.03888130933046341,
+ "learning_rate": 0.0005988281427877498,
+ "loss": 2.915215015411377,
+ "step": 200
+ },
+ {
+ "epoch": 2.794425087108014,
+ "grad_norm": 0.037759196013212204,
+ "learning_rate": 0.000598804365316586,
+ "loss": 2.9155189990997314,
+ "step": 201
+ },
+ {
+ "epoch": 2.808362369337979,
+ "grad_norm": 0.03914877027273178,
+ "learning_rate": 0.0005987803495211879,
+ "loss": 2.9149580001831055,
+ "step": 202
+ },
+ {
+ "epoch": 2.822299651567944,
+ "grad_norm": 0.03494696691632271,
+ "learning_rate": 0.0005987560954209047,
+ "loss": 2.9019455909729004,
+ "step": 203
+ },
+ {
+ "epoch": 2.8362369337979096,
+ "grad_norm": 0.04843524843454361,
+ "learning_rate": 0.0005987316030352773,
+ "loss": 2.906419038772583,
+ "step": 204
+ },
+ {
+ "epoch": 2.8501742160278747,
+ "grad_norm": 0.04746173322200775,
+ "learning_rate": 0.000598706872384039,
+ "loss": 2.90865421295166,
+ "step": 205
+ },
+ {
+ "epoch": 2.8641114982578397,
+ "grad_norm": 0.07141132652759552,
+ "learning_rate": 0.0005986819034871147,
+ "loss": 2.911919116973877,
+ "step": 206
+ },
+ {
+ "epoch": 2.8780487804878048,
+ "grad_norm": 0.07191284745931625,
+ "learning_rate": 0.0005986566963646214,
+ "loss": 2.896134614944458,
+ "step": 207
+ },
+ {
+ "epoch": 2.89198606271777,
+ "grad_norm": 0.054023921489715576,
+ "learning_rate": 0.000598631251036868,
+ "loss": 2.8979387283325195,
+ "step": 208
+ },
+ {
+ "epoch": 2.9059233449477353,
+ "grad_norm": 0.05115993320941925,
+ "learning_rate": 0.0005986055675243555,
+ "loss": 2.895747661590576,
+ "step": 209
+ },
+ {
+ "epoch": 2.9198606271777003,
+ "grad_norm": 0.055029138922691345,
+ "learning_rate": 0.0005985796458477765,
+ "loss": 2.901477336883545,
+ "step": 210
+ },
+ {
+ "epoch": 2.9337979094076654,
+ "grad_norm": 0.050139401108026505,
+ "learning_rate": 0.0005985534860280155,
+ "loss": 2.899230480194092,
+ "step": 211
+ },
+ {
+ "epoch": 2.947735191637631,
+ "grad_norm": 0.04797767847776413,
+ "learning_rate": 0.0005985270880861493,
+ "loss": 2.8821463584899902,
+ "step": 212
+ },
+ {
+ "epoch": 2.961672473867596,
+ "grad_norm": 0.032600875943899155,
+ "learning_rate": 0.0005985004520434462,
+ "loss": 2.876875400543213,
+ "step": 213
+ },
+ {
+ "epoch": 2.975609756097561,
+ "grad_norm": 0.037163764238357544,
+ "learning_rate": 0.000598473577921366,
+ "loss": 2.8857016563415527,
+ "step": 214
+ },
+ {
+ "epoch": 2.989547038327526,
+ "grad_norm": 0.03918052464723587,
+ "learning_rate": 0.0005984464657415611,
+ "loss": 2.8761796951293945,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.0412505678832531,
+ "learning_rate": 0.0005984191155258751,
+ "loss": 2.160161018371582,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 0.724179208278656,
+ "eval_runtime": 40.6035,
+ "eval_samples_per_second": 60.143,
+ "eval_steps_per_second": 0.493,
+ "step": 216
+ },
+ {
+ "epoch": 3.013937282229965,
+ "grad_norm": 0.03409483656287193,
+ "learning_rate": 0.0005983915272963436,
+ "loss": 2.871135711669922,
+ "step": 217
+ },
+ {
+ "epoch": 3.0278745644599305,
+ "grad_norm": 0.04408197104930878,
+ "learning_rate": 0.0005983637010751941,
+ "loss": 2.867302417755127,
+ "step": 218
+ },
+ {
+ "epoch": 3.0418118466898956,
+ "grad_norm": 0.06749062240123749,
+ "learning_rate": 0.0005983356368848455,
+ "loss": 2.871088981628418,
+ "step": 219
+ },
+ {
+ "epoch": 3.0557491289198606,
+ "grad_norm": 0.10093795508146286,
+ "learning_rate": 0.0005983073347479086,
+ "loss": 2.878399133682251,
+ "step": 220
+ },
+ {
+ "epoch": 3.0696864111498257,
+ "grad_norm": 0.10228322446346283,
+ "learning_rate": 0.000598278794687186,
+ "loss": 2.8688411712646484,
+ "step": 221
+ },
+ {
+ "epoch": 3.083623693379791,
+ "grad_norm": 0.08215560019016266,
+ "learning_rate": 0.0005982500167256719,
+ "loss": 2.872134208679199,
+ "step": 222
+ },
+ {
+ "epoch": 3.097560975609756,
+ "grad_norm": 0.08236275613307953,
+ "learning_rate": 0.0005982210008865522,
+ "loss": 2.870941162109375,
+ "step": 223
+ },
+ {
+ "epoch": 3.1114982578397212,
+ "grad_norm": 0.07016550004482269,
+ "learning_rate": 0.0005981917471932045,
+ "loss": 2.870575428009033,
+ "step": 224
+ },
+ {
+ "epoch": 3.1254355400696863,
+ "grad_norm": 0.07636428624391556,
+ "learning_rate": 0.0005981622556691978,
+ "loss": 2.868832588195801,
+ "step": 225
+ },
+ {
+ "epoch": 3.1393728222996518,
+ "grad_norm": 0.08546213805675507,
+ "learning_rate": 0.0005981325263382931,
+ "loss": 2.869138240814209,
+ "step": 226
+ },
+ {
+ "epoch": 3.153310104529617,
+ "grad_norm": 0.0976535901427269,
+ "learning_rate": 0.0005981025592244425,
+ "loss": 2.867882251739502,
+ "step": 227
+ },
+ {
+ "epoch": 3.167247386759582,
+ "grad_norm": 0.09015904366970062,
+ "learning_rate": 0.00059807235435179,
+ "loss": 2.8630735874176025,
+ "step": 228
+ },
+ {
+ "epoch": 3.181184668989547,
+ "grad_norm": 0.0658399909734726,
+ "learning_rate": 0.0005980419117446715,
+ "loss": 2.8700802326202393,
+ "step": 229
+ },
+ {
+ "epoch": 3.1951219512195124,
+ "grad_norm": 0.05341155081987381,
+ "learning_rate": 0.0005980112314276138,
+ "loss": 2.8630599975585938,
+ "step": 230
+ },
+ {
+ "epoch": 3.2090592334494774,
+ "grad_norm": 0.05429157242178917,
+ "learning_rate": 0.0005979803134253354,
+ "loss": 2.8516578674316406,
+ "step": 231
+ },
+ {
+ "epoch": 3.2229965156794425,
+ "grad_norm": 0.043862055987119675,
+ "learning_rate": 0.0005979491577627464,
+ "loss": 2.8573832511901855,
+ "step": 232
+ },
+ {
+ "epoch": 3.2369337979094075,
+ "grad_norm": 0.048795755952596664,
+ "learning_rate": 0.0005979177644649485,
+ "loss": 2.853893280029297,
+ "step": 233
+ },
+ {
+ "epoch": 3.250871080139373,
+ "grad_norm": 0.03508317843079567,
+ "learning_rate": 0.0005978861335572346,
+ "loss": 2.8559699058532715,
+ "step": 234
+ },
+ {
+ "epoch": 3.264808362369338,
+ "grad_norm": 0.039702750742435455,
+ "learning_rate": 0.0005978542650650892,
+ "loss": 2.8468103408813477,
+ "step": 235
+ },
+ {
+ "epoch": 3.278745644599303,
+ "grad_norm": 0.03869146108627319,
+ "learning_rate": 0.0005978221590141881,
+ "loss": 2.8342695236206055,
+ "step": 236
+ },
+ {
+ "epoch": 3.292682926829268,
+ "grad_norm": 0.03934340551495552,
+ "learning_rate": 0.0005977898154303987,
+ "loss": 2.8470571041107178,
+ "step": 237
+ },
+ {
+ "epoch": 3.3066202090592336,
+ "grad_norm": 0.037353284657001495,
+ "learning_rate": 0.0005977572343397794,
+ "loss": 2.840059518814087,
+ "step": 238
+ },
+ {
+ "epoch": 3.3205574912891986,
+ "grad_norm": 0.032541144639253616,
+ "learning_rate": 0.0005977244157685805,
+ "loss": 2.847280979156494,
+ "step": 239
+ },
+ {
+ "epoch": 3.3344947735191637,
+ "grad_norm": 0.03452054038643837,
+ "learning_rate": 0.000597691359743243,
+ "loss": 2.8502321243286133,
+ "step": 240
+ },
+ {
+ "epoch": 3.3484320557491287,
+ "grad_norm": 0.03700891137123108,
+ "learning_rate": 0.0005976580662903999,
+ "loss": 2.83978533744812,
+ "step": 241
+ },
+ {
+ "epoch": 3.362369337979094,
+ "grad_norm": 0.029760025441646576,
+ "learning_rate": 0.0005976245354368749,
+ "loss": 2.8311290740966797,
+ "step": 242
+ },
+ {
+ "epoch": 3.3763066202090593,
+ "grad_norm": 0.03517364338040352,
+ "learning_rate": 0.0005975907672096832,
+ "loss": 2.824399948120117,
+ "step": 243
+ },
+ {
+ "epoch": 3.3902439024390243,
+ "grad_norm": 0.04522453993558884,
+ "learning_rate": 0.0005975567616360313,
+ "loss": 2.8246240615844727,
+ "step": 244
+ },
+ {
+ "epoch": 3.40418118466899,
+ "grad_norm": 0.0494384802877903,
+ "learning_rate": 0.0005975225187433169,
+ "loss": 2.8449745178222656,
+ "step": 245
+ },
+ {
+ "epoch": 3.418118466898955,
+ "grad_norm": 0.061527837067842484,
+ "learning_rate": 0.000597488038559129,
+ "loss": 2.8286070823669434,
+ "step": 246
+ },
+ {
+ "epoch": 3.43205574912892,
+ "grad_norm": 0.07200496643781662,
+ "learning_rate": 0.0005974533211112474,
+ "loss": 2.83577823638916,
+ "step": 247
+ },
+ {
+ "epoch": 3.445993031358885,
+ "grad_norm": 0.08873147517442703,
+ "learning_rate": 0.0005974183664276434,
+ "loss": 2.836007595062256,
+ "step": 248
+ },
+ {
+ "epoch": 3.45993031358885,
+ "grad_norm": 0.09570424258708954,
+ "learning_rate": 0.0005973831745364795,
+ "loss": 2.840940237045288,
+ "step": 249
+ },
+ {
+ "epoch": 3.4738675958188154,
+ "grad_norm": 0.10362695902585983,
+ "learning_rate": 0.000597347745466109,
+ "loss": 2.850740909576416,
+ "step": 250
+ },
+ {
+ "epoch": 3.4878048780487805,
+ "grad_norm": 0.08370107412338257,
+ "learning_rate": 0.0005973120792450766,
+ "loss": 2.8572049140930176,
+ "step": 251
+ },
+ {
+ "epoch": 3.5017421602787455,
+ "grad_norm": 0.0737074464559555,
+ "learning_rate": 0.0005972761759021178,
+ "loss": 2.8149335384368896,
+ "step": 252
+ },
+ {
+ "epoch": 3.515679442508711,
+ "grad_norm": 0.07435688376426697,
+ "learning_rate": 0.0005972400354661594,
+ "loss": 2.817427158355713,
+ "step": 253
+ },
+ {
+ "epoch": 3.529616724738676,
+ "grad_norm": 0.07407844066619873,
+ "learning_rate": 0.000597203657966319,
+ "loss": 2.8169541358947754,
+ "step": 254
+ },
+ {
+ "epoch": 3.543554006968641,
+ "grad_norm": 0.07459145039319992,
+ "learning_rate": 0.0005971670434319053,
+ "loss": 2.8235528469085693,
+ "step": 255
+ },
+ {
+ "epoch": 3.557491289198606,
+ "grad_norm": 0.0899617001414299,
+ "learning_rate": 0.0005971301918924182,
+ "loss": 2.832531690597534,
+ "step": 256
+ },
+ {
+ "epoch": 3.571428571428571,
+ "grad_norm": 0.09097135066986084,
+ "learning_rate": 0.0005970931033775479,
+ "loss": 2.8258728981018066,
+ "step": 257
+ },
+ {
+ "epoch": 3.5853658536585367,
+ "grad_norm": 0.07991395890712738,
+ "learning_rate": 0.0005970557779171763,
+ "loss": 2.8159890174865723,
+ "step": 258
+ },
+ {
+ "epoch": 3.5993031358885017,
+ "grad_norm": 0.0683487132191658,
+ "learning_rate": 0.0005970182155413756,
+ "loss": 2.819324016571045,
+ "step": 259
+ },
+ {
+ "epoch": 3.6132404181184667,
+ "grad_norm": 0.05294633284211159,
+ "learning_rate": 0.0005969804162804093,
+ "loss": 2.8166258335113525,
+ "step": 260
+ },
+ {
+ "epoch": 3.6271777003484322,
+ "grad_norm": 0.04694054648280144,
+ "learning_rate": 0.0005969423801647314,
+ "loss": 2.813283681869507,
+ "step": 261
+ },
+ {
+ "epoch": 3.6411149825783973,
+ "grad_norm": 0.04447995498776436,
+ "learning_rate": 0.0005969041072249872,
+ "loss": 2.8058133125305176,
+ "step": 262
+ },
+ {
+ "epoch": 3.6550522648083623,
+ "grad_norm": 0.03876696154475212,
+ "learning_rate": 0.0005968655974920124,
+ "loss": 2.8110270500183105,
+ "step": 263
+ },
+ {
+ "epoch": 3.6689895470383274,
+ "grad_norm": 0.0417216531932354,
+ "learning_rate": 0.0005968268509968335,
+ "loss": 2.8145129680633545,
+ "step": 264
+ },
+ {
+ "epoch": 3.682926829268293,
+ "grad_norm": 0.03208528086543083,
+ "learning_rate": 0.0005967878677706678,
+ "loss": 2.7957448959350586,
+ "step": 265
+ },
+ {
+ "epoch": 3.696864111498258,
+ "grad_norm": 0.04042236506938934,
+ "learning_rate": 0.0005967486478449236,
+ "loss": 2.7897958755493164,
+ "step": 266
+ },
+ {
+ "epoch": 3.710801393728223,
+ "grad_norm": 0.038808777928352356,
+ "learning_rate": 0.0005967091912511996,
+ "loss": 2.8041954040527344,
+ "step": 267
+ },
+ {
+ "epoch": 3.7247386759581884,
+ "grad_norm": 0.037022914737463,
+ "learning_rate": 0.0005966694980212851,
+ "loss": 2.809565782546997,
+ "step": 268
+ },
+ {
+ "epoch": 3.7386759581881535,
+ "grad_norm": 0.0356953889131546,
+ "learning_rate": 0.0005966295681871604,
+ "loss": 2.794809341430664,
+ "step": 269
+ },
+ {
+ "epoch": 3.7526132404181185,
+ "grad_norm": 0.03807281702756882,
+ "learning_rate": 0.0005965894017809962,
+ "loss": 2.795717239379883,
+ "step": 270
+ },
+ {
+ "epoch": 3.7665505226480835,
+ "grad_norm": 0.042434707283973694,
+ "learning_rate": 0.0005965489988351539,
+ "loss": 2.8004324436187744,
+ "step": 271
+ },
+ {
+ "epoch": 3.7804878048780486,
+ "grad_norm": 0.0522489920258522,
+ "learning_rate": 0.0005965083593821853,
+ "loss": 2.790482521057129,
+ "step": 272
+ },
+ {
+ "epoch": 3.794425087108014,
+ "grad_norm": 0.06727278232574463,
+ "learning_rate": 0.000596467483454833,
+ "loss": 2.7881433963775635,
+ "step": 273
+ },
+ {
+ "epoch": 3.808362369337979,
+ "grad_norm": 0.08363860845565796,
+ "learning_rate": 0.0005964263710860299,
+ "loss": 2.7925806045532227,
+ "step": 274
+ },
+ {
+ "epoch": 3.822299651567944,
+ "grad_norm": 0.06598818302154541,
+ "learning_rate": 0.0005963850223088995,
+ "loss": 2.772061347961426,
+ "step": 275
+ },
+ {
+ "epoch": 3.8362369337979096,
+ "grad_norm": 0.054081548005342484,
+ "learning_rate": 0.0005963434371567559,
+ "loss": 2.7749743461608887,
+ "step": 276
+ },
+ {
+ "epoch": 3.8501742160278747,
+ "grad_norm": 0.05758000165224075,
+ "learning_rate": 0.0005963016156631033,
+ "loss": 2.7791082859039307,
+ "step": 277
+ },
+ {
+ "epoch": 3.8641114982578397,
+ "grad_norm": 0.050822217017412186,
+ "learning_rate": 0.0005962595578616367,
+ "loss": 2.7815473079681396,
+ "step": 278
+ },
+ {
+ "epoch": 3.8780487804878048,
+ "grad_norm": 0.040895845741033554,
+ "learning_rate": 0.0005962172637862411,
+ "loss": 2.773261547088623,
+ "step": 279
+ },
+ {
+ "epoch": 3.89198606271777,
+ "grad_norm": 0.034904301166534424,
+ "learning_rate": 0.0005961747334709924,
+ "loss": 2.7627172470092773,
+ "step": 280
+ },
+ {
+ "epoch": 3.9059233449477353,
+ "grad_norm": 0.041809309273958206,
+ "learning_rate": 0.0005961319669501562,
+ "loss": 2.758397102355957,
+ "step": 281
+ },
+ {
+ "epoch": 3.9198606271777003,
+ "grad_norm": 0.03494604676961899,
+ "learning_rate": 0.0005960889642581887,
+ "loss": 2.776466131210327,
+ "step": 282
+ },
+ {
+ "epoch": 3.9337979094076654,
+ "grad_norm": 0.03635790944099426,
+ "learning_rate": 0.0005960457254297367,
+ "loss": 2.7763757705688477,
+ "step": 283
+ },
+ {
+ "epoch": 3.947735191637631,
+ "grad_norm": 0.035400789231061935,
+ "learning_rate": 0.0005960022504996369,
+ "loss": 2.759999990463257,
+ "step": 284
+ },
+ {
+ "epoch": 3.961672473867596,
+ "grad_norm": 0.04523325711488724,
+ "learning_rate": 0.0005959585395029158,
+ "loss": 2.7523179054260254,
+ "step": 285
+ },
+ {
+ "epoch": 3.975609756097561,
+ "grad_norm": 0.05560452491044998,
+ "learning_rate": 0.000595914592474791,
+ "loss": 2.758845567703247,
+ "step": 286
+ },
+ {
+ "epoch": 3.989547038327526,
+ "grad_norm": 0.06153145059943199,
+ "learning_rate": 0.0005958704094506698,
+ "loss": 2.749399185180664,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.05208462104201317,
+ "learning_rate": 0.0005958259904661497,
+ "loss": 2.071669340133667,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 0.695522665977478,
+ "eval_runtime": 40.3189,
+ "eval_samples_per_second": 60.567,
+ "eval_steps_per_second": 0.496,
+ "step": 288
+ },
+ {
+ "epoch": 4.013937282229965,
+ "grad_norm": 0.05128208547830582,
+ "learning_rate": 0.0005957813355570178,
+ "loss": 2.739234447479248,
+ "step": 289
+ },
+ {
+ "epoch": 4.02787456445993,
+ "grad_norm": 0.07697609066963196,
+ "learning_rate": 0.0005957364447592524,
+ "loss": 2.7523460388183594,
+ "step": 290
+ },
+ {
+ "epoch": 4.041811846689895,
+ "grad_norm": 0.07936502993106842,
+ "learning_rate": 0.0005956913181090208,
+ "loss": 2.7405805587768555,
+ "step": 291
+ },
+ {
+ "epoch": 4.055749128919861,
+ "grad_norm": 0.07429715991020203,
+ "learning_rate": 0.0005956459556426809,
+ "loss": 2.760530710220337,
+ "step": 292
+ },
+ {
+ "epoch": 4.069686411149826,
+ "grad_norm": 0.06815112382173538,
+ "learning_rate": 0.0005956003573967802,
+ "loss": 2.7342095375061035,
+ "step": 293
+ },
+ {
+ "epoch": 4.083623693379791,
+ "grad_norm": 0.07373050600290298,
+ "learning_rate": 0.0005955545234080567,
+ "loss": 2.7389070987701416,
+ "step": 294
+ },
+ {
+ "epoch": 4.097560975609756,
+ "grad_norm": 0.08975639194250107,
+ "learning_rate": 0.0005955084537134378,
+ "loss": 2.756620168685913,
+ "step": 295
+ },
+ {
+ "epoch": 4.111498257839721,
+ "grad_norm": 0.11307737231254578,
+ "learning_rate": 0.0005954621483500411,
+ "loss": 2.757890224456787,
+ "step": 296
+ },
+ {
+ "epoch": 4.125435540069686,
+ "grad_norm": 0.10326922684907913,
+ "learning_rate": 0.0005954156073551739,
+ "loss": 2.7604823112487793,
+ "step": 297
+ },
+ {
+ "epoch": 4.139372822299651,
+ "grad_norm": 0.07577540725469589,
+ "learning_rate": 0.0005953688307663336,
+ "loss": 2.761768341064453,
+ "step": 298
+ },
+ {
+ "epoch": 4.153310104529616,
+ "grad_norm": 0.06620045751333237,
+ "learning_rate": 0.0005953218186212072,
+ "loss": 2.748669147491455,
+ "step": 299
+ },
+ {
+ "epoch": 4.167247386759582,
+ "grad_norm": 0.06167765334248543,
+ "learning_rate": 0.0005952745709576714,
+ "loss": 2.741281509399414,
+ "step": 300
+ },
+ {
+ "epoch": 4.181184668989547,
+ "grad_norm": 0.06262649595737457,
+ "learning_rate": 0.000595227087813793,
+ "loss": 2.7453837394714355,
+ "step": 301
+ },
+ {
+ "epoch": 4.195121951219512,
+ "grad_norm": 0.07164547592401505,
+ "learning_rate": 0.0005951793692278282,
+ "loss": 2.757172107696533,
+ "step": 302
+ },
+ {
+ "epoch": 4.209059233449477,
+ "grad_norm": 0.06647909432649612,
+ "learning_rate": 0.0005951314152382229,
+ "loss": 2.7389321327209473,
+ "step": 303
+ },
+ {
+ "epoch": 4.2229965156794425,
+ "grad_norm": 0.061594728380441666,
+ "learning_rate": 0.0005950832258836129,
+ "loss": 2.738513946533203,
+ "step": 304
+ },
+ {
+ "epoch": 4.2369337979094075,
+ "grad_norm": 0.06701034307479858,
+ "learning_rate": 0.0005950348012028235,
+ "loss": 2.7303874492645264,
+ "step": 305
+ },
+ {
+ "epoch": 4.2508710801393725,
+ "grad_norm": 0.06630577892065048,
+ "learning_rate": 0.0005949861412348694,
+ "loss": 2.737123966217041,
+ "step": 306
+ },
+ {
+ "epoch": 4.264808362369338,
+ "grad_norm": 0.04944160208106041,
+ "learning_rate": 0.0005949372460189555,
+ "loss": 2.7435293197631836,
+ "step": 307
+ },
+ {
+ "epoch": 4.2787456445993035,
+ "grad_norm": 0.04269982501864433,
+ "learning_rate": 0.0005948881155944753,
+ "loss": 2.72158145904541,
+ "step": 308
+ },
+ {
+ "epoch": 4.2926829268292686,
+ "grad_norm": 0.05561396852135658,
+ "learning_rate": 0.0005948387500010126,
+ "loss": 2.731154203414917,
+ "step": 309
+ },
+ {
+ "epoch": 4.306620209059234,
+ "grad_norm": 0.049912456423044205,
+ "learning_rate": 0.0005947891492783401,
+ "loss": 2.7142367362976074,
+ "step": 310
+ },
+ {
+ "epoch": 4.320557491289199,
+ "grad_norm": 0.04867382347583771,
+ "learning_rate": 0.0005947393134664205,
+ "loss": 2.7216851711273193,
+ "step": 311
+ },
+ {
+ "epoch": 4.334494773519164,
+ "grad_norm": 0.05156482011079788,
+ "learning_rate": 0.0005946892426054054,
+ "loss": 2.717257261276245,
+ "step": 312
+ },
+ {
+ "epoch": 4.348432055749129,
+ "grad_norm": 0.05031168460845947,
+ "learning_rate": 0.0005946389367356361,
+ "loss": 2.7293102741241455,
+ "step": 313
+ },
+ {
+ "epoch": 4.362369337979094,
+ "grad_norm": 0.03850070387125015,
+ "learning_rate": 0.0005945883958976432,
+ "loss": 2.720397472381592,
+ "step": 314
+ },
+ {
+ "epoch": 4.376306620209059,
+ "grad_norm": 0.049057621508836746,
+ "learning_rate": 0.0005945376201321464,
+ "loss": 2.71873140335083,
+ "step": 315
+ },
+ {
+ "epoch": 4.390243902439025,
+ "grad_norm": 0.039578285068273544,
+ "learning_rate": 0.0005944866094800548,
+ "loss": 2.726874351501465,
+ "step": 316
+ },
+ {
+ "epoch": 4.40418118466899,
+ "grad_norm": 0.03636026754975319,
+ "learning_rate": 0.0005944353639824669,
+ "loss": 2.7162535190582275,
+ "step": 317
+ },
+ {
+ "epoch": 4.418118466898955,
+ "grad_norm": 0.03454010188579559,
+ "learning_rate": 0.0005943838836806702,
+ "loss": 2.7009294033050537,
+ "step": 318
+ },
+ {
+ "epoch": 4.43205574912892,
+ "grad_norm": 0.032978180795907974,
+ "learning_rate": 0.0005943321686161414,
+ "loss": 2.709333896636963,
+ "step": 319
+ },
+ {
+ "epoch": 4.445993031358885,
+ "grad_norm": 0.03197623789310455,
+ "learning_rate": 0.0005942802188305464,
+ "loss": 2.702411413192749,
+ "step": 320
+ },
+ {
+ "epoch": 4.45993031358885,
+ "grad_norm": 0.02986901067197323,
+ "learning_rate": 0.0005942280343657403,
+ "loss": 2.6926870346069336,
+ "step": 321
+ },
+ {
+ "epoch": 4.473867595818815,
+ "grad_norm": 0.03517908602952957,
+ "learning_rate": 0.0005941756152637671,
+ "loss": 2.6933770179748535,
+ "step": 322
+ },
+ {
+ "epoch": 4.487804878048781,
+ "grad_norm": 0.03140024468302727,
+ "learning_rate": 0.00059412296156686,
+ "loss": 2.710171937942505,
+ "step": 323
+ },
+ {
+ "epoch": 4.501742160278746,
+ "grad_norm": 0.03267952427268028,
+ "learning_rate": 0.0005940700733174409,
+ "loss": 2.69620418548584,
+ "step": 324
+ },
+ {
+ "epoch": 4.515679442508711,
+ "grad_norm": 0.03277324512600899,
+ "learning_rate": 0.0005940169505581213,
+ "loss": 2.6958892345428467,
+ "step": 325
+ },
+ {
+ "epoch": 4.529616724738676,
+ "grad_norm": 0.03559919074177742,
+ "learning_rate": 0.0005939635933317009,
+ "loss": 2.6927599906921387,
+ "step": 326
+ },
+ {
+ "epoch": 4.543554006968641,
+ "grad_norm": 0.03539760038256645,
+ "learning_rate": 0.0005939100016811688,
+ "loss": 2.6677207946777344,
+ "step": 327
+ },
+ {
+ "epoch": 4.557491289198606,
+ "grad_norm": 0.04007532075047493,
+ "learning_rate": 0.0005938561756497028,
+ "loss": 2.687222480773926,
+ "step": 328
+ },
+ {
+ "epoch": 4.571428571428571,
+ "grad_norm": 0.03828917816281319,
+ "learning_rate": 0.0005938021152806696,
+ "loss": 2.6915855407714844,
+ "step": 329
+ },
+ {
+ "epoch": 4.585365853658536,
+ "grad_norm": 0.043163347989320755,
+ "learning_rate": 0.0005937478206176248,
+ "loss": 2.683889389038086,
+ "step": 330
+ },
+ {
+ "epoch": 4.599303135888501,
+ "grad_norm": 0.05287671089172363,
+ "learning_rate": 0.0005936932917043124,
+ "loss": 2.696169853210449,
+ "step": 331
+ },
+ {
+ "epoch": 4.613240418118467,
+ "grad_norm": 0.06691489368677139,
+ "learning_rate": 0.0005936385285846651,
+ "loss": 2.6903016567230225,
+ "step": 332
+ },
+ {
+ "epoch": 4.627177700348432,
+ "grad_norm": 0.08349309861660004,
+ "learning_rate": 0.0005935835313028052,
+ "loss": 2.6813721656799316,
+ "step": 333
+ },
+ {
+ "epoch": 4.641114982578397,
+ "grad_norm": 0.09731950610876083,
+ "learning_rate": 0.0005935282999030426,
+ "loss": 2.691209554672241,
+ "step": 334
+ },
+ {
+ "epoch": 4.655052264808362,
+ "grad_norm": 0.07833823561668396,
+ "learning_rate": 0.0005934728344298763,
+ "loss": 2.6876442432403564,
+ "step": 335
+ },
+ {
+ "epoch": 4.668989547038327,
+ "grad_norm": 0.0696968138217926,
+ "learning_rate": 0.000593417134927994,
+ "loss": 2.6932926177978516,
+ "step": 336
+ },
+ {
+ "epoch": 4.682926829268292,
+ "grad_norm": 0.0632258951663971,
+ "learning_rate": 0.0005933612014422715,
+ "loss": 2.6817989349365234,
+ "step": 337
+ },
+ {
+ "epoch": 4.696864111498257,
+ "grad_norm": 0.0748044028878212,
+ "learning_rate": 0.0005933050340177736,
+ "loss": 2.6821188926696777,
+ "step": 338
+ },
+ {
+ "epoch": 4.710801393728223,
+ "grad_norm": 0.06642459332942963,
+ "learning_rate": 0.0005932486326997534,
+ "loss": 2.6929712295532227,
+ "step": 339
+ },
+ {
+ "epoch": 4.724738675958188,
+ "grad_norm": 0.06058414652943611,
+ "learning_rate": 0.0005931919975336523,
+ "loss": 2.692411422729492,
+ "step": 340
+ },
+ {
+ "epoch": 4.7386759581881535,
+ "grad_norm": 0.05587534233927727,
+ "learning_rate": 0.0005931351285651004,
+ "loss": 2.676719903945923,
+ "step": 341
+ },
+ {
+ "epoch": 4.7526132404181185,
+ "grad_norm": 0.059625860303640366,
+ "learning_rate": 0.0005930780258399157,
+ "loss": 2.6882309913635254,
+ "step": 342
+ },
+ {
+ "epoch": 4.7665505226480835,
+ "grad_norm": 0.05370939150452614,
+ "learning_rate": 0.0005930206894041051,
+ "loss": 2.6796631813049316,
+ "step": 343
+ },
+ {
+ "epoch": 4.780487804878049,
+ "grad_norm": 0.06609658896923065,
+ "learning_rate": 0.0005929631193038634,
+ "loss": 2.687939405441284,
+ "step": 344
+ },
+ {
+ "epoch": 4.794425087108014,
+ "grad_norm": 0.07027886062860489,
+ "learning_rate": 0.0005929053155855739,
+ "loss": 2.672358989715576,
+ "step": 345
+ },
+ {
+ "epoch": 4.80836236933798,
+ "grad_norm": 0.05532550811767578,
+ "learning_rate": 0.0005928472782958079,
+ "loss": 2.6738951206207275,
+ "step": 346
+ },
+ {
+ "epoch": 4.822299651567945,
+ "grad_norm": 0.04593169689178467,
+ "learning_rate": 0.000592789007481325,
+ "loss": 2.675957679748535,
+ "step": 347
+ },
+ {
+ "epoch": 4.83623693379791,
+ "grad_norm": 0.04651476442813873,
+ "learning_rate": 0.0005927305031890732,
+ "loss": 2.6599249839782715,
+ "step": 348
+ },
+ {
+ "epoch": 4.850174216027875,
+ "grad_norm": 0.04131711646914482,
+ "learning_rate": 0.000592671765466188,
+ "loss": 2.672339916229248,
+ "step": 349
+ },
+ {
+ "epoch": 4.86411149825784,
+ "grad_norm": 0.03916105255484581,
+ "learning_rate": 0.0005926127943599934,
+ "loss": 2.654345989227295,
+ "step": 350
+ },
+ {
+ "epoch": 4.878048780487805,
+ "grad_norm": 0.04122985526919365,
+ "learning_rate": 0.0005925535899180015,
+ "loss": 2.6729297637939453,
+ "step": 351
+ },
+ {
+ "epoch": 4.89198606271777,
+ "grad_norm": 0.058476220816373825,
+ "learning_rate": 0.0005924941521879122,
+ "loss": 2.665987014770508,
+ "step": 352
+ },
+ {
+ "epoch": 4.905923344947735,
+ "grad_norm": 0.07334969192743301,
+ "learning_rate": 0.0005924344812176134,
+ "loss": 2.661764144897461,
+ "step": 353
+ },
+ {
+ "epoch": 4.9198606271777,
+ "grad_norm": 0.05559230595827103,
+ "learning_rate": 0.0005923745770551808,
+ "loss": 2.6589086055755615,
+ "step": 354
+ },
+ {
+ "epoch": 4.933797909407666,
+ "grad_norm": 0.04733807221055031,
+ "learning_rate": 0.0005923144397488782,
+ "loss": 2.6471903324127197,
+ "step": 355
+ },
+ {
+ "epoch": 4.947735191637631,
+ "grad_norm": 0.056338120251894,
+ "learning_rate": 0.0005922540693471572,
+ "loss": 2.66196346282959,
+ "step": 356
+ },
+ {
+ "epoch": 4.961672473867596,
+ "grad_norm": 0.070899099111557,
+ "learning_rate": 0.0005921934658986571,
+ "loss": 2.6613497734069824,
+ "step": 357
+ },
+ {
+ "epoch": 4.975609756097561,
+ "grad_norm": 0.05829300358891487,
+ "learning_rate": 0.000592132629452205,
+ "loss": 2.6537318229675293,
+ "step": 358
+ },
+ {
+ "epoch": 4.989547038327526,
+ "grad_norm": 0.04470338672399521,
+ "learning_rate": 0.0005920715600568156,
+ "loss": 2.639193296432495,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.029524177312850952,
+ "learning_rate": 0.0005920102577616916,
+ "loss": 1.9856886863708496,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 0.6703579425811768,
+ "eval_runtime": 40.1947,
+ "eval_samples_per_second": 60.754,
+ "eval_steps_per_second": 0.498,
+ "step": 360
+ },
+ {
+ "epoch": 5.013937282229965,
+ "grad_norm": 0.039727699011564255,
+ "learning_rate": 0.0005919487226162231,
+ "loss": 2.6248669624328613,
+ "step": 361
+ },
+ {
+ "epoch": 5.02787456445993,
+ "grad_norm": 0.039338793605566025,
+ "learning_rate": 0.0005918869546699877,
+ "loss": 2.630005359649658,
+ "step": 362
+ },
+ {
+ "epoch": 5.041811846689895,
+ "grad_norm": 0.037072643637657166,
+ "learning_rate": 0.0005918249539727509,
+ "loss": 2.6241390705108643,
+ "step": 363
+ },
+ {
+ "epoch": 5.055749128919861,
+ "grad_norm": 0.035198889672756195,
+ "learning_rate": 0.0005917627205744655,
+ "loss": 2.6221063137054443,
+ "step": 364
+ },
+ {
+ "epoch": 5.069686411149826,
+ "grad_norm": 0.04578937590122223,
+ "learning_rate": 0.0005917002545252719,
+ "loss": 2.6351819038391113,
+ "step": 365
+ },
+ {
+ "epoch": 5.083623693379791,
+ "grad_norm": 0.0523366779088974,
+ "learning_rate": 0.0005916375558754977,
+ "loss": 2.62751841545105,
+ "step": 366
+ },
+ {
+ "epoch": 5.097560975609756,
+ "grad_norm": 0.060026299208402634,
+ "learning_rate": 0.0005915746246756581,
+ "loss": 2.625197410583496,
+ "step": 367
+ },
+ {
+ "epoch": 5.111498257839721,
+ "grad_norm": 0.05982828885316849,
+ "learning_rate": 0.0005915114609764558,
+ "loss": 2.619140148162842,
+ "step": 368
+ },
+ {
+ "epoch": 5.125435540069686,
+ "grad_norm": 0.04769567400217056,
+ "learning_rate": 0.0005914480648287804,
+ "loss": 2.6197752952575684,
+ "step": 369
+ },
+ {
+ "epoch": 5.139372822299651,
+ "grad_norm": 0.04090019315481186,
+ "learning_rate": 0.0005913844362837093,
+ "loss": 2.6253864765167236,
+ "step": 370
+ },
+ {
+ "epoch": 5.153310104529616,
+ "grad_norm": 0.04769155755639076,
+ "learning_rate": 0.0005913205753925066,
+ "loss": 2.6180930137634277,
+ "step": 371
+ },
+ {
+ "epoch": 5.167247386759582,
+ "grad_norm": 0.048125989735126495,
+ "learning_rate": 0.0005912564822066241,
+ "loss": 2.613513708114624,
+ "step": 372
+ },
+ {
+ "epoch": 5.181184668989547,
+ "grad_norm": 0.049780264496803284,
+ "learning_rate": 0.0005911921567777005,
+ "loss": 2.6197900772094727,
+ "step": 373
+ },
+ {
+ "epoch": 5.195121951219512,
+ "grad_norm": 0.04975919425487518,
+ "learning_rate": 0.0005911275991575613,
+ "loss": 2.6189403533935547,
+ "step": 374
+ },
+ {
+ "epoch": 5.209059233449477,
+ "grad_norm": 0.045773062855005264,
+ "learning_rate": 0.0005910628093982198,
+ "loss": 2.6092190742492676,
+ "step": 375
+ },
+ {
+ "epoch": 5.2229965156794425,
+ "grad_norm": 0.041725803166627884,
+ "learning_rate": 0.0005909977875518759,
+ "loss": 2.6096010208129883,
+ "step": 376
+ },
+ {
+ "epoch": 5.2369337979094075,
+ "grad_norm": 0.04649988189339638,
+ "learning_rate": 0.0005909325336709164,
+ "loss": 2.6014528274536133,
+ "step": 377
+ },
+ {
+ "epoch": 5.2508710801393725,
+ "grad_norm": 0.05227779224514961,
+ "learning_rate": 0.0005908670478079152,
+ "loss": 2.604328155517578,
+ "step": 378
+ },
+ {
+ "epoch": 5.264808362369338,
+ "grad_norm": 0.04857666417956352,
+ "learning_rate": 0.000590801330015633,
+ "loss": 2.6103639602661133,
+ "step": 379
+ },
+ {
+ "epoch": 5.2787456445993035,
+ "grad_norm": 0.047866497188806534,
+ "learning_rate": 0.0005907353803470177,
+ "loss": 2.6128387451171875,
+ "step": 380
+ },
+ {
+ "epoch": 5.2926829268292686,
+ "grad_norm": 0.05275033041834831,
+ "learning_rate": 0.0005906691988552034,
+ "loss": 2.6159093379974365,
+ "step": 381
+ },
+ {
+ "epoch": 5.306620209059234,
+ "grad_norm": 0.06210039183497429,
+ "learning_rate": 0.0005906027855935115,
+ "loss": 2.6041178703308105,
+ "step": 382
+ },
+ {
+ "epoch": 5.320557491289199,
+ "grad_norm": 0.06392814964056015,
+ "learning_rate": 0.0005905361406154501,
+ "loss": 2.6114046573638916,
+ "step": 383
+ },
+ {
+ "epoch": 5.334494773519164,
+ "grad_norm": 0.0644841268658638,
+ "learning_rate": 0.0005904692639747137,
+ "loss": 2.6141104698181152,
+ "step": 384
+ },
+ {
+ "epoch": 5.348432055749129,
+ "grad_norm": 0.06638580560684204,
+ "learning_rate": 0.0005904021557251837,
+ "loss": 2.616746187210083,
+ "step": 385
+ },
+ {
+ "epoch": 5.362369337979094,
+ "grad_norm": 0.06176018714904785,
+ "learning_rate": 0.0005903348159209277,
+ "loss": 2.6172876358032227,
+ "step": 386
+ },
+ {
+ "epoch": 5.376306620209059,
+ "grad_norm": 0.052066221833229065,
+ "learning_rate": 0.0005902672446162007,
+ "loss": 2.5987703800201416,
+ "step": 387
+ },
+ {
+ "epoch": 5.390243902439025,
+ "grad_norm": 0.047028761357069016,
+ "learning_rate": 0.0005901994418654432,
+ "loss": 2.596569538116455,
+ "step": 388
+ },
+ {
+ "epoch": 5.40418118466899,
+ "grad_norm": 0.05026818439364433,
+ "learning_rate": 0.0005901314077232829,
+ "loss": 2.5997366905212402,
+ "step": 389
+ },
+ {
+ "epoch": 5.418118466898955,
+ "grad_norm": 0.04762030020356178,
+ "learning_rate": 0.0005900631422445335,
+ "loss": 2.598598003387451,
+ "step": 390
+ },
+ {
+ "epoch": 5.43205574912892,
+ "grad_norm": 0.05125858634710312,
+ "learning_rate": 0.0005899946454841955,
+ "loss": 2.605160713195801,
+ "step": 391
+ },
+ {
+ "epoch": 5.445993031358885,
+ "grad_norm": 0.054070550948381424,
+ "learning_rate": 0.0005899259174974553,
+ "loss": 2.6062188148498535,
+ "step": 392
+ },
+ {
+ "epoch": 5.45993031358885,
+ "grad_norm": 0.0688333809375763,
+ "learning_rate": 0.000589856958339686,
+ "loss": 2.5885558128356934,
+ "step": 393
+ },
+ {
+ "epoch": 5.473867595818815,
+ "grad_norm": 0.07510587573051453,
+ "learning_rate": 0.0005897877680664465,
+ "loss": 2.5977489948272705,
+ "step": 394
+ },
+ {
+ "epoch": 5.487804878048781,
+ "grad_norm": 0.05772198736667633,
+ "learning_rate": 0.0005897183467334822,
+ "loss": 2.6046319007873535,
+ "step": 395
+ },
+ {
+ "epoch": 5.501742160278746,
+ "grad_norm": 0.05309263989329338,
+ "learning_rate": 0.0005896486943967249,
+ "loss": 2.593275785446167,
+ "step": 396
+ },
+ {
+ "epoch": 5.515679442508711,
+ "grad_norm": 0.05616075545549393,
+ "learning_rate": 0.0005895788111122919,
+ "loss": 2.568248748779297,
+ "step": 397
+ },
+ {
+ "epoch": 5.529616724738676,
+ "grad_norm": 0.04955418035387993,
+ "learning_rate": 0.0005895086969364871,
+ "loss": 2.5889053344726562,
+ "step": 398
+ },
+ {
+ "epoch": 5.543554006968641,
+ "grad_norm": 0.04541563615202904,
+ "learning_rate": 0.0005894383519258001,
+ "loss": 2.6085779666900635,
+ "step": 399
+ },
+ {
+ "epoch": 5.557491289198606,
+ "grad_norm": 0.052929993718862534,
+ "learning_rate": 0.0005893677761369066,
+ "loss": 2.5928592681884766,
+ "step": 400
+ },
+ {
+ "epoch": 5.571428571428571,
+ "grad_norm": 0.05284162238240242,
+ "learning_rate": 0.0005892969696266683,
+ "loss": 2.572826862335205,
+ "step": 401
+ },
+ {
+ "epoch": 5.585365853658536,
+ "grad_norm": 0.0423022136092186,
+ "learning_rate": 0.0005892259324521328,
+ "loss": 2.5721311569213867,
+ "step": 402
+ },
+ {
+ "epoch": 5.599303135888501,
+ "grad_norm": 0.039413321763277054,
+ "learning_rate": 0.0005891546646705334,
+ "loss": 2.574954032897949,
+ "step": 403
+ },
+ {
+ "epoch": 5.613240418118467,
+ "grad_norm": 0.03749777376651764,
+ "learning_rate": 0.0005890831663392891,
+ "loss": 2.580699920654297,
+ "step": 404
+ },
+ {
+ "epoch": 5.627177700348432,
+ "grad_norm": 0.037166696041822433,
+ "learning_rate": 0.0005890114375160052,
+ "loss": 2.5765609741210938,
+ "step": 405
+ },
+ {
+ "epoch": 5.641114982578397,
+ "grad_norm": 0.034370459616184235,
+ "learning_rate": 0.0005889394782584718,
+ "loss": 2.588606357574463,
+ "step": 406
+ },
+ {
+ "epoch": 5.655052264808362,
+ "grad_norm": 0.03365951031446457,
+ "learning_rate": 0.0005888672886246656,
+ "loss": 2.57226824760437,
+ "step": 407
+ },
+ {
+ "epoch": 5.668989547038327,
+ "grad_norm": 0.03125793859362602,
+ "learning_rate": 0.0005887948686727483,
+ "loss": 2.5793633460998535,
+ "step": 408
+ },
+ {
+ "epoch": 5.682926829268292,
+ "grad_norm": 0.027411725372076035,
+ "learning_rate": 0.0005887222184610675,
+ "loss": 2.5830397605895996,
+ "step": 409
+ },
+ {
+ "epoch": 5.696864111498257,
+ "grad_norm": 0.028178710490465164,
+ "learning_rate": 0.0005886493380481559,
+ "loss": 2.5680036544799805,
+ "step": 410
+ },
+ {
+ "epoch": 5.710801393728223,
+ "grad_norm": 0.02988891676068306,
+ "learning_rate": 0.0005885762274927322,
+ "loss": 2.573624849319458,
+ "step": 411
+ },
+ {
+ "epoch": 5.724738675958188,
+ "grad_norm": 0.02810891903936863,
+ "learning_rate": 0.0005885028868537,
+ "loss": 2.582005023956299,
+ "step": 412
+ },
+ {
+ "epoch": 5.7386759581881535,
+ "grad_norm": 0.031402505934238434,
+ "learning_rate": 0.0005884293161901487,
+ "loss": 2.5672507286071777,
+ "step": 413
+ },
+ {
+ "epoch": 5.7526132404181185,
+ "grad_norm": 0.03285510838031769,
+ "learning_rate": 0.000588355515561353,
+ "loss": 2.571962356567383,
+ "step": 414
+ },
+ {
+ "epoch": 5.7665505226480835,
+ "grad_norm": 0.03331591933965683,
+ "learning_rate": 0.0005882814850267723,
+ "loss": 2.573042392730713,
+ "step": 415
+ },
+ {
+ "epoch": 5.780487804878049,
+ "grad_norm": 0.035851649940013885,
+ "learning_rate": 0.0005882072246460521,
+ "loss": 2.582005023956299,
+ "step": 416
+ },
+ {
+ "epoch": 5.794425087108014,
+ "grad_norm": 0.04279213026165962,
+ "learning_rate": 0.0005881327344790223,
+ "loss": 2.5464048385620117,
+ "step": 417
+ },
+ {
+ "epoch": 5.80836236933798,
+ "grad_norm": 0.04937538132071495,
+ "learning_rate": 0.0005880580145856986,
+ "loss": 2.558198928833008,
+ "step": 418
+ },
+ {
+ "epoch": 5.822299651567945,
+ "grad_norm": 0.04781151935458183,
+ "learning_rate": 0.0005879830650262813,
+ "loss": 2.571126937866211,
+ "step": 419
+ },
+ {
+ "epoch": 5.83623693379791,
+ "grad_norm": 0.04407578706741333,
+ "learning_rate": 0.0005879078858611557,
+ "loss": 2.5579514503479004,
+ "step": 420
+ },
+ {
+ "epoch": 5.850174216027875,
+ "grad_norm": 0.04472474008798599,
+ "learning_rate": 0.0005878324771508927,
+ "loss": 2.57476806640625,
+ "step": 421
+ },
+ {
+ "epoch": 5.86411149825784,
+ "grad_norm": 0.039548397064208984,
+ "learning_rate": 0.0005877568389562475,
+ "loss": 2.5661427974700928,
+ "step": 422
+ },
+ {
+ "epoch": 5.878048780487805,
+ "grad_norm": 0.04089987650513649,
+ "learning_rate": 0.0005876809713381606,
+ "loss": 2.55307936668396,
+ "step": 423
+ },
+ {
+ "epoch": 5.89198606271777,
+ "grad_norm": 0.05161741375923157,
+ "learning_rate": 0.0005876048743577569,
+ "loss": 2.564471960067749,
+ "step": 424
+ },
+ {
+ "epoch": 5.905923344947735,
+ "grad_norm": 0.057691119611263275,
+ "learning_rate": 0.0005875285480763466,
+ "loss": 2.5518410205841064,
+ "step": 425
+ },
+ {
+ "epoch": 5.9198606271777,
+ "grad_norm": 0.05954357609152794,
+ "learning_rate": 0.0005874519925554244,
+ "loss": 2.5482420921325684,
+ "step": 426
+ },
+ {
+ "epoch": 5.933797909407666,
+ "grad_norm": 0.06078341230750084,
+ "learning_rate": 0.0005873752078566694,
+ "loss": 2.559572696685791,
+ "step": 427
+ },
+ {
+ "epoch": 5.947735191637631,
+ "grad_norm": 0.060291387140750885,
+ "learning_rate": 0.0005872981940419459,
+ "loss": 2.5643410682678223,
+ "step": 428
+ },
+ {
+ "epoch": 5.961672473867596,
+ "grad_norm": 0.06805447489023209,
+ "learning_rate": 0.0005872209511733027,
+ "loss": 2.5565760135650635,
+ "step": 429
+ },
+ {
+ "epoch": 5.975609756097561,
+ "grad_norm": 0.08052453398704529,
+ "learning_rate": 0.0005871434793129726,
+ "loss": 2.5634193420410156,
+ "step": 430
+ },
+ {
+ "epoch": 5.989547038327526,
+ "grad_norm": 0.07786231487989426,
+ "learning_rate": 0.0005870657785233735,
+ "loss": 2.5665931701660156,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.057144131511449814,
+ "learning_rate": 0.0005869878488671075,
+ "loss": 1.9077787399291992,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 0.6502015590667725,
+ "eval_runtime": 40.5144,
+ "eval_samples_per_second": 60.275,
+ "eval_steps_per_second": 0.494,
+ "step": 432
+ },
+ {
+ "epoch": 6.013937282229965,
+ "grad_norm": 0.062045276165008545,
+ "learning_rate": 0.0005869096904069611,
+ "loss": 2.5416250228881836,
+ "step": 433
+ },
+ {
+ "epoch": 6.02787456445993,
+ "grad_norm": 0.0648002102971077,
+ "learning_rate": 0.0005868313032059052,
+ "loss": 2.5361738204956055,
+ "step": 434
+ },
+ {
+ "epoch": 6.041811846689895,
+ "grad_norm": 0.06482209265232086,
+ "learning_rate": 0.0005867526873270949,
+ "loss": 2.54353404045105,
+ "step": 435
+ },
+ {
+ "epoch": 6.055749128919861,
+ "grad_norm": 0.05858788639307022,
+ "learning_rate": 0.0005866738428338695,
+ "loss": 2.5441808700561523,
+ "step": 436
+ },
+ {
+ "epoch": 6.069686411149826,
+ "grad_norm": 0.05710245296359062,
+ "learning_rate": 0.000586594769789753,
+ "loss": 2.5357842445373535,
+ "step": 437
+ },
+ {
+ "epoch": 6.083623693379791,
+ "grad_norm": 0.05553581565618515,
+ "learning_rate": 0.0005865154682584524,
+ "loss": 2.5358052253723145,
+ "step": 438
+ },
+ {
+ "epoch": 6.097560975609756,
+ "grad_norm": 0.055729176849126816,
+ "learning_rate": 0.0005864359383038602,
+ "loss": 2.5209150314331055,
+ "step": 439
+ },
+ {
+ "epoch": 6.111498257839721,
+ "grad_norm": 0.04890104755759239,
+ "learning_rate": 0.000586356179990052,
+ "loss": 2.52427339553833,
+ "step": 440
+ },
+ {
+ "epoch": 6.125435540069686,
+ "grad_norm": 0.056426040828228,
+ "learning_rate": 0.0005862761933812875,
+ "loss": 2.5387511253356934,
+ "step": 441
+ },
+ {
+ "epoch": 6.139372822299651,
+ "grad_norm": 0.059780314564704895,
+ "learning_rate": 0.0005861959785420106,
+ "loss": 2.531538724899292,
+ "step": 442
+ },
+ {
+ "epoch": 6.153310104529616,
+ "grad_norm": 0.05493956059217453,
+ "learning_rate": 0.000586115535536849,
+ "loss": 2.5271177291870117,
+ "step": 443
+ },
+ {
+ "epoch": 6.167247386759582,
+ "grad_norm": 0.05240267142653465,
+ "learning_rate": 0.000586034864430614,
+ "loss": 2.533730983734131,
+ "step": 444
+ },
+ {
+ "epoch": 6.181184668989547,
+ "grad_norm": 0.057911988347768784,
+ "learning_rate": 0.0005859539652883008,
+ "loss": 2.532154083251953,
+ "step": 445
+ },
+ {
+ "epoch": 6.195121951219512,
+ "grad_norm": 0.06957171857357025,
+ "learning_rate": 0.0005858728381750888,
+ "loss": 2.5321333408355713,
+ "step": 446
+ },
+ {
+ "epoch": 6.209059233449477,
+ "grad_norm": 0.055091492831707,
+ "learning_rate": 0.0005857914831563402,
+ "loss": 2.5275330543518066,
+ "step": 447
+ },
+ {
+ "epoch": 6.2229965156794425,
+ "grad_norm": 0.04599588364362717,
+ "learning_rate": 0.0005857099002976016,
+ "loss": 2.5294699668884277,
+ "step": 448
+ },
+ {
+ "epoch": 6.2369337979094075,
+ "grad_norm": 0.05306335911154747,
+ "learning_rate": 0.0005856280896646026,
+ "loss": 2.535618782043457,
+ "step": 449
+ },
+ {
+ "epoch": 6.2508710801393725,
+ "grad_norm": 0.05369280278682709,
+ "learning_rate": 0.0005855460513232567,
+ "loss": 2.5181729793548584,
+ "step": 450
+ },
+ {
+ "epoch": 6.264808362369338,
+ "grad_norm": 0.04584018141031265,
+ "learning_rate": 0.0005854637853396606,
+ "loss": 2.5095596313476562,
+ "step": 451
+ },
+ {
+ "epoch": 6.2787456445993035,
+ "grad_norm": 0.045887988060712814,
+ "learning_rate": 0.0005853812917800945,
+ "loss": 2.543792247772217,
+ "step": 452
+ },
+ {
+ "epoch": 6.2926829268292686,
+ "grad_norm": 0.04554528370499611,
+ "learning_rate": 0.0005852985707110221,
+ "loss": 2.514249086380005,
+ "step": 453
+ },
+ {
+ "epoch": 6.306620209059234,
+ "grad_norm": 0.044292885810136795,
+ "learning_rate": 0.0005852156221990901,
+ "loss": 2.521683931350708,
+ "step": 454
+ },
+ {
+ "epoch": 6.320557491289199,
+ "grad_norm": 0.04753238335251808,
+ "learning_rate": 0.0005851324463111289,
+ "loss": 2.5285496711730957,
+ "step": 455
+ },
+ {
+ "epoch": 6.334494773519164,
+ "grad_norm": 0.053385961800813675,
+ "learning_rate": 0.0005850490431141516,
+ "loss": 2.5080885887145996,
+ "step": 456
+ },
+ {
+ "epoch": 6.348432055749129,
+ "grad_norm": 0.043983977288007736,
+ "learning_rate": 0.0005849654126753545,
+ "loss": 2.5061187744140625,
+ "step": 457
+ },
+ {
+ "epoch": 6.362369337979094,
+ "grad_norm": 0.04052867367863655,
+ "learning_rate": 0.0005848815550621175,
+ "loss": 2.5316476821899414,
+ "step": 458
+ },
+ {
+ "epoch": 6.376306620209059,
+ "grad_norm": 0.03497035801410675,
+ "learning_rate": 0.0005847974703420028,
+ "loss": 2.517639636993408,
+ "step": 459
+ },
+ {
+ "epoch": 6.390243902439025,
+ "grad_norm": 0.03455311805009842,
+ "learning_rate": 0.0005847131585827564,
+ "loss": 2.517422914505005,
+ "step": 460
+ },
+ {
+ "epoch": 6.40418118466899,
+ "grad_norm": 0.040654607117176056,
+ "learning_rate": 0.0005846286198523061,
+ "loss": 2.506796360015869,
+ "step": 461
+ },
+ {
+ "epoch": 6.418118466898955,
+ "grad_norm": 0.03895020857453346,
+ "learning_rate": 0.0005845438542187638,
+ "loss": 2.4978837966918945,
+ "step": 462
+ },
+ {
+ "epoch": 6.43205574912892,
+ "grad_norm": 0.03841928392648697,
+ "learning_rate": 0.0005844588617504236,
+ "loss": 2.512120246887207,
+ "step": 463
+ },
+ {
+ "epoch": 6.445993031358885,
+ "grad_norm": 0.03369898721575737,
+ "learning_rate": 0.0005843736425157621,
+ "loss": 2.500955820083618,
+ "step": 464
+ },
+ {
+ "epoch": 6.45993031358885,
+ "grad_norm": 0.03319835662841797,
+ "learning_rate": 0.000584288196583439,
+ "loss": 2.495816707611084,
+ "step": 465
+ },
+ {
+ "epoch": 6.473867595818815,
+ "grad_norm": 0.03281501680612564,
+ "learning_rate": 0.0005842025240222966,
+ "loss": 2.4982664585113525,
+ "step": 466
+ },
+ {
+ "epoch": 6.487804878048781,
+ "grad_norm": 0.035637401044368744,
+ "learning_rate": 0.0005841166249013598,
+ "loss": 2.5070347785949707,
+ "step": 467
+ },
+ {
+ "epoch": 6.501742160278746,
+ "grad_norm": 0.031222257763147354,
+ "learning_rate": 0.0005840304992898359,
+ "loss": 2.5108470916748047,
+ "step": 468
+ },
+ {
+ "epoch": 6.515679442508711,
+ "grad_norm": 0.03228495270013809,
+ "learning_rate": 0.0005839441472571147,
+ "loss": 2.4994547367095947,
+ "step": 469
+ },
+ {
+ "epoch": 6.529616724738676,
+ "grad_norm": 0.0315462164580822,
+ "learning_rate": 0.0005838575688727685,
+ "loss": 2.517374038696289,
+ "step": 470
+ },
+ {
+ "epoch": 6.543554006968641,
+ "grad_norm": 0.03304170072078705,
+ "learning_rate": 0.0005837707642065518,
+ "loss": 2.504943370819092,
+ "step": 471
+ },
+ {
+ "epoch": 6.557491289198606,
+ "grad_norm": 0.03496484458446503,
+ "learning_rate": 0.000583683733328402,
+ "loss": 2.509988784790039,
+ "step": 472
+ },
+ {
+ "epoch": 6.571428571428571,
+ "grad_norm": 0.04144975170493126,
+ "learning_rate": 0.0005835964763084378,
+ "loss": 2.4944121837615967,
+ "step": 473
+ },
+ {
+ "epoch": 6.585365853658536,
+ "grad_norm": 0.04626752808690071,
+ "learning_rate": 0.000583508993216961,
+ "loss": 2.519324541091919,
+ "step": 474
+ },
+ {
+ "epoch": 6.599303135888501,
+ "grad_norm": 0.045151665806770325,
+ "learning_rate": 0.0005834212841244548,
+ "loss": 2.5022945404052734,
+ "step": 475
+ },
+ {
+ "epoch": 6.613240418118467,
+ "grad_norm": 0.04149537533521652,
+ "learning_rate": 0.0005833333491015851,
+ "loss": 2.496169328689575,
+ "step": 476
+ },
+ {
+ "epoch": 6.627177700348432,
+ "grad_norm": 0.0439605787396431,
+ "learning_rate": 0.0005832451882191995,
+ "loss": 2.503659248352051,
+ "step": 477
+ },
+ {
+ "epoch": 6.641114982578397,
+ "grad_norm": 0.041153572499752045,
+ "learning_rate": 0.0005831568015483274,
+ "loss": 2.508594036102295,
+ "step": 478
+ },
+ {
+ "epoch": 6.655052264808362,
+ "grad_norm": 0.03570788353681564,
+ "learning_rate": 0.0005830681891601807,
+ "loss": 2.5032777786254883,
+ "step": 479
+ },
+ {
+ "epoch": 6.668989547038327,
+ "grad_norm": 0.042115096002817154,
+ "learning_rate": 0.0005829793511261525,
+ "loss": 2.4948456287384033,
+ "step": 480
+ },
+ {
+ "epoch": 6.682926829268292,
+ "grad_norm": 0.04520317539572716,
+ "learning_rate": 0.0005828902875178183,
+ "loss": 2.5003604888916016,
+ "step": 481
+ },
+ {
+ "epoch": 6.696864111498257,
+ "grad_norm": 0.045228149741888046,
+ "learning_rate": 0.0005828009984069347,
+ "loss": 2.503664493560791,
+ "step": 482
+ },
+ {
+ "epoch": 6.710801393728223,
+ "grad_norm": 0.051417600363492966,
+ "learning_rate": 0.0005827114838654406,
+ "loss": 2.488755226135254,
+ "step": 483
+ },
+ {
+ "epoch": 6.724738675958188,
+ "grad_norm": 0.049600034952163696,
+ "learning_rate": 0.0005826217439654562,
+ "loss": 2.4962430000305176,
+ "step": 484
+ },
+ {
+ "epoch": 6.7386759581881535,
+ "grad_norm": 0.044539641588926315,
+ "learning_rate": 0.0005825317787792831,
+ "loss": 2.508113384246826,
+ "step": 485
+ },
+ {
+ "epoch": 6.7526132404181185,
+ "grad_norm": 0.04709797352552414,
+ "learning_rate": 0.0005824415883794049,
+ "loss": 2.5054588317871094,
+ "step": 486
+ },
+ {
+ "epoch": 6.7665505226480835,
+ "grad_norm": 0.05492638051509857,
+ "learning_rate": 0.0005823511728384863,
+ "loss": 2.4963388442993164,
+ "step": 487
+ },
+ {
+ "epoch": 6.780487804878049,
+ "grad_norm": 0.061837874352931976,
+ "learning_rate": 0.0005822605322293733,
+ "loss": 2.4929351806640625,
+ "step": 488
+ },
+ {
+ "epoch": 6.794425087108014,
+ "grad_norm": 0.06532683968544006,
+ "learning_rate": 0.0005821696666250937,
+ "loss": 2.4878501892089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.80836236933798,
+ "grad_norm": 0.05990716814994812,
+ "learning_rate": 0.0005820785760988559,
+ "loss": 2.496943473815918,
+ "step": 490
+ },
+ {
+ "epoch": 6.822299651567945,
+ "grad_norm": 0.04542370140552521,
+ "learning_rate": 0.0005819872607240503,
+ "loss": 2.5100255012512207,
+ "step": 491
+ },
+ {
+ "epoch": 6.83623693379791,
+ "grad_norm": 0.04482433572411537,
+ "learning_rate": 0.0005818957205742478,
+ "loss": 2.4926881790161133,
+ "step": 492
+ },
+ {
+ "epoch": 6.850174216027875,
+ "grad_norm": 0.04078618437051773,
+ "learning_rate": 0.0005818039557232005,
+ "loss": 2.5003790855407715,
+ "step": 493
+ },
+ {
+ "epoch": 6.86411149825784,
+ "grad_norm": 0.04109889268875122,
+ "learning_rate": 0.0005817119662448421,
+ "loss": 2.512845039367676,
+ "step": 494
+ },
+ {
+ "epoch": 6.878048780487805,
+ "grad_norm": 0.03949440270662308,
+ "learning_rate": 0.0005816197522132866,
+ "loss": 2.488072395324707,
+ "step": 495
+ },
+ {
+ "epoch": 6.89198606271777,
+ "grad_norm": 0.040081024169921875,
+ "learning_rate": 0.0005815273137028292,
+ "loss": 2.504167079925537,
+ "step": 496
+ },
+ {
+ "epoch": 6.905923344947735,
+ "grad_norm": 0.03655780106782913,
+ "learning_rate": 0.0005814346507879459,
+ "loss": 2.4962496757507324,
+ "step": 497
+ },
+ {
+ "epoch": 6.9198606271777,
+ "grad_norm": 0.03624923527240753,
+ "learning_rate": 0.0005813417635432937,
+ "loss": 2.4843900203704834,
+ "step": 498
+ },
+ {
+ "epoch": 6.933797909407666,
+ "grad_norm": 0.034214284271001816,
+ "learning_rate": 0.00058124865204371,
+ "loss": 2.4778432846069336,
+ "step": 499
+ },
+ {
+ "epoch": 6.947735191637631,
+ "grad_norm": 0.0354473851621151,
+ "learning_rate": 0.0005811553163642131,
+ "loss": 2.49300217628479,
+ "step": 500
+ },
+ {
+ "epoch": 6.961672473867596,
+ "grad_norm": 0.04188355430960655,
+ "learning_rate": 0.0005810617565800019,
+ "loss": 2.4916625022888184,
+ "step": 501
+ },
+ {
+ "epoch": 6.975609756097561,
+ "grad_norm": 0.044823627918958664,
+ "learning_rate": 0.0005809679727664559,
+ "loss": 2.502657651901245,
+ "step": 502
+ },
+ {
+ "epoch": 6.989547038327526,
+ "grad_norm": 0.0468236543238163,
+ "learning_rate": 0.0005808739649991348,
+ "loss": 2.473053455352783,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.03717072308063507,
+ "learning_rate": 0.0005807797333537792,
+ "loss": 1.8738106489181519,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 0.6349480748176575,
+ "eval_runtime": 39.8506,
+ "eval_samples_per_second": 61.279,
+ "eval_steps_per_second": 0.502,
+ "step": 504
+ },
+ {
+ "epoch": 7.013937282229965,
+ "grad_norm": 0.0408112071454525,
+ "learning_rate": 0.0005806852779063098,
+ "loss": 2.446186065673828,
+ "step": 505
+ },
+ {
+ "epoch": 7.02787456445993,
+ "grad_norm": 0.046356331557035446,
+ "learning_rate": 0.0005805905987328275,
+ "loss": 2.46285343170166,
+ "step": 506
+ },
+ {
+ "epoch": 7.041811846689895,
+ "grad_norm": 0.0413004569709301,
+ "learning_rate": 0.0005804956959096138,
+ "loss": 2.4553096294403076,
+ "step": 507
+ },
+ {
+ "epoch": 7.055749128919861,
+ "grad_norm": 0.04176029935479164,
+ "learning_rate": 0.0005804005695131299,
+ "loss": 2.4663143157958984,
+ "step": 508
+ },
+ {
+ "epoch": 7.069686411149826,
+ "grad_norm": 0.04505695775151253,
+ "learning_rate": 0.0005803052196200177,
+ "loss": 2.4530575275421143,
+ "step": 509
+ },
+ {
+ "epoch": 7.083623693379791,
+ "grad_norm": 0.05135089531540871,
+ "learning_rate": 0.0005802096463070988,
+ "loss": 2.4337000846862793,
+ "step": 510
+ },
+ {
+ "epoch": 7.097560975609756,
+ "grad_norm": 0.06322603672742844,
+ "learning_rate": 0.0005801138496513749,
+ "loss": 2.4787750244140625,
+ "step": 511
+ },
+ {
+ "epoch": 7.111498257839721,
+ "grad_norm": 0.0680660754442215,
+ "learning_rate": 0.0005800178297300276,
+ "loss": 2.462891101837158,
+ "step": 512
+ },
+ {
+ "epoch": 7.125435540069686,
+ "grad_norm": 0.07223042100667953,
+ "learning_rate": 0.0005799215866204183,
+ "loss": 2.466489553451538,
+ "step": 513
+ },
+ {
+ "epoch": 7.139372822299651,
+ "grad_norm": 0.0868474468588829,
+ "learning_rate": 0.0005798251204000886,
+ "loss": 2.4692301750183105,
+ "step": 514
+ },
+ {
+ "epoch": 7.153310104529616,
+ "grad_norm": 0.07377298176288605,
+ "learning_rate": 0.0005797284311467594,
+ "loss": 2.4755783081054688,
+ "step": 515
+ },
+ {
+ "epoch": 7.167247386759582,
+ "grad_norm": 0.07302770018577576,
+ "learning_rate": 0.0005796315189383316,
+ "loss": 2.4619765281677246,
+ "step": 516
+ },
+ {
+ "epoch": 7.181184668989547,
+ "grad_norm": 0.06365817785263062,
+ "learning_rate": 0.0005795343838528855,
+ "loss": 2.469318151473999,
+ "step": 517
+ },
+ {
+ "epoch": 7.195121951219512,
+ "grad_norm": 0.06436990201473236,
+ "learning_rate": 0.0005794370259686811,
+ "loss": 2.4581692218780518,
+ "step": 518
+ },
+ {
+ "epoch": 7.209059233449477,
+ "grad_norm": 0.06494493037462234,
+ "learning_rate": 0.0005793394453641581,
+ "loss": 2.469299793243408,
+ "step": 519
+ },
+ {
+ "epoch": 7.2229965156794425,
+ "grad_norm": 0.07596383988857269,
+ "learning_rate": 0.0005792416421179351,
+ "loss": 2.4772355556488037,
+ "step": 520
+ },
+ {
+ "epoch": 7.2369337979094075,
+ "grad_norm": 0.07409915328025818,
+ "learning_rate": 0.0005791436163088106,
+ "loss": 2.477510929107666,
+ "step": 521
+ },
+ {
+ "epoch": 7.2508710801393725,
+ "grad_norm": 0.06815416365861893,
+ "learning_rate": 0.0005790453680157622,
+ "loss": 2.4639534950256348,
+ "step": 522
+ },
+ {
+ "epoch": 7.264808362369338,
+ "grad_norm": 0.06958118826150894,
+ "learning_rate": 0.0005789468973179468,
+ "loss": 2.4708127975463867,
+ "step": 523
+ },
+ {
+ "epoch": 7.2787456445993035,
+ "grad_norm": 0.06792039424180984,
+ "learning_rate": 0.0005788482042947004,
+ "loss": 2.462704658508301,
+ "step": 524
+ },
+ {
+ "epoch": 7.2926829268292686,
+ "grad_norm": 0.058236561715602875,
+ "learning_rate": 0.0005787492890255382,
+ "loss": 2.455707550048828,
+ "step": 525
+ },
+ {
+ "epoch": 7.306620209059234,
+ "grad_norm": 0.05740198865532875,
+ "learning_rate": 0.0005786501515901545,
+ "loss": 2.472905158996582,
+ "step": 526
+ },
+ {
+ "epoch": 7.320557491289199,
+ "grad_norm": 0.05712047219276428,
+ "learning_rate": 0.0005785507920684225,
+ "loss": 2.4714770317077637,
+ "step": 527
+ },
+ {
+ "epoch": 7.334494773519164,
+ "grad_norm": 0.05431913211941719,
+ "learning_rate": 0.0005784512105403944,
+ "loss": 2.467907190322876,
+ "step": 528
+ },
+ {
+ "epoch": 7.348432055749129,
+ "grad_norm": 0.05013139173388481,
+ "learning_rate": 0.000578351407086301,
+ "loss": 2.4667627811431885,
+ "step": 529
+ },
+ {
+ "epoch": 7.362369337979094,
+ "grad_norm": 0.04926477372646332,
+ "learning_rate": 0.0005782513817865527,
+ "loss": 2.4545669555664062,
+ "step": 530
+ },
+ {
+ "epoch": 7.376306620209059,
+ "grad_norm": 0.04078412428498268,
+ "learning_rate": 0.0005781511347217375,
+ "loss": 2.4482979774475098,
+ "step": 531
+ },
+ {
+ "epoch": 7.390243902439025,
+ "grad_norm": 0.03570006415247917,
+ "learning_rate": 0.000578050665972623,
+ "loss": 2.4588446617126465,
+ "step": 532
+ },
+ {
+ "epoch": 7.40418118466899,
+ "grad_norm": 0.04253046214580536,
+ "learning_rate": 0.0005779499756201549,
+ "loss": 2.463991641998291,
+ "step": 533
+ },
+ {
+ "epoch": 7.418118466898955,
+ "grad_norm": 0.03903041407465935,
+ "learning_rate": 0.0005778490637454576,
+ "loss": 2.4524803161621094,
+ "step": 534
+ },
+ {
+ "epoch": 7.43205574912892,
+ "grad_norm": 0.036041416227817535,
+ "learning_rate": 0.000577747930429834,
+ "loss": 2.466279983520508,
+ "step": 535
+ },
+ {
+ "epoch": 7.445993031358885,
+ "grad_norm": 0.036224234849214554,
+ "learning_rate": 0.0005776465757547654,
+ "loss": 2.444584369659424,
+ "step": 536
+ },
+ {
+ "epoch": 7.45993031358885,
+ "grad_norm": 0.03627604618668556,
+ "learning_rate": 0.0005775449998019113,
+ "loss": 2.4629859924316406,
+ "step": 537
+ },
+ {
+ "epoch": 7.473867595818815,
+ "grad_norm": 0.03370306268334389,
+ "learning_rate": 0.00057744320265311,
+ "loss": 2.448319435119629,
+ "step": 538
+ },
+ {
+ "epoch": 7.487804878048781,
+ "grad_norm": 0.031237779185175896,
+ "learning_rate": 0.0005773411843903773,
+ "loss": 2.4637656211853027,
+ "step": 539
+ },
+ {
+ "epoch": 7.501742160278746,
+ "grad_norm": 0.03511526808142662,
+ "learning_rate": 0.0005772389450959075,
+ "loss": 2.4473280906677246,
+ "step": 540
+ },
+ {
+ "epoch": 7.515679442508711,
+ "grad_norm": 0.03426751121878624,
+ "learning_rate": 0.000577136484852073,
+ "loss": 2.443058490753174,
+ "step": 541
+ },
+ {
+ "epoch": 7.529616724738676,
+ "grad_norm": 0.03289111703634262,
+ "learning_rate": 0.000577033803741424,
+ "loss": 2.448587417602539,
+ "step": 542
+ },
+ {
+ "epoch": 7.543554006968641,
+ "grad_norm": 0.03324120491743088,
+ "learning_rate": 0.0005769309018466891,
+ "loss": 2.448965549468994,
+ "step": 543
+ },
+ {
+ "epoch": 7.557491289198606,
+ "grad_norm": 0.031128088012337685,
+ "learning_rate": 0.0005768277792507744,
+ "loss": 2.4602575302124023,
+ "step": 544
+ },
+ {
+ "epoch": 7.571428571428571,
+ "grad_norm": 0.033455077558755875,
+ "learning_rate": 0.0005767244360367638,
+ "loss": 2.4488325119018555,
+ "step": 545
+ },
+ {
+ "epoch": 7.585365853658536,
+ "grad_norm": 0.0313849113881588,
+ "learning_rate": 0.0005766208722879192,
+ "loss": 2.4347963333129883,
+ "step": 546
+ },
+ {
+ "epoch": 7.599303135888501,
+ "grad_norm": 0.03434115648269653,
+ "learning_rate": 0.00057651708808768,
+ "loss": 2.4518353939056396,
+ "step": 547
+ },
+ {
+ "epoch": 7.613240418118467,
+ "grad_norm": 0.03155781701207161,
+ "learning_rate": 0.0005764130835196632,
+ "loss": 2.443643808364868,
+ "step": 548
+ },
+ {
+ "epoch": 7.627177700348432,
+ "grad_norm": 0.028192076832056046,
+ "learning_rate": 0.0005763088586676634,
+ "loss": 2.451793670654297,
+ "step": 549
+ },
+ {
+ "epoch": 7.641114982578397,
+ "grad_norm": 0.03205734118819237,
+ "learning_rate": 0.0005762044136156528,
+ "loss": 2.4506397247314453,
+ "step": 550
+ },
+ {
+ "epoch": 7.655052264808362,
+ "grad_norm": 0.03486161306500435,
+ "learning_rate": 0.0005760997484477809,
+ "loss": 2.455686569213867,
+ "step": 551
+ },
+ {
+ "epoch": 7.668989547038327,
+ "grad_norm": 0.02914281003177166,
+ "learning_rate": 0.0005759948632483742,
+ "loss": 2.443521022796631,
+ "step": 552
+ },
+ {
+ "epoch": 7.682926829268292,
+ "grad_norm": 0.03087310679256916,
+ "learning_rate": 0.0005758897581019371,
+ "loss": 2.455596923828125,
+ "step": 553
+ },
+ {
+ "epoch": 7.696864111498257,
+ "grad_norm": 0.032375626266002655,
+ "learning_rate": 0.000575784433093151,
+ "loss": 2.436648368835449,
+ "step": 554
+ },
+ {
+ "epoch": 7.710801393728223,
+ "grad_norm": 0.031863704323768616,
+ "learning_rate": 0.0005756788883068743,
+ "loss": 2.447986125946045,
+ "step": 555
+ },
+ {
+ "epoch": 7.724738675958188,
+ "grad_norm": 0.03159843757748604,
+ "learning_rate": 0.0005755731238281423,
+ "loss": 2.4244322776794434,
+ "step": 556
+ },
+ {
+ "epoch": 7.7386759581881535,
+ "grad_norm": 0.03639158979058266,
+ "learning_rate": 0.0005754671397421678,
+ "loss": 2.438507556915283,
+ "step": 557
+ },
+ {
+ "epoch": 7.7526132404181185,
+ "grad_norm": 0.035257723182439804,
+ "learning_rate": 0.0005753609361343402,
+ "loss": 2.437483310699463,
+ "step": 558
+ },
+ {
+ "epoch": 7.7665505226480835,
+ "grad_norm": 0.034112557768821716,
+ "learning_rate": 0.0005752545130902256,
+ "loss": 2.4336776733398438,
+ "step": 559
+ },
+ {
+ "epoch": 7.780487804878049,
+ "grad_norm": 0.03131738305091858,
+ "learning_rate": 0.0005751478706955674,
+ "loss": 2.4558157920837402,
+ "step": 560
+ },
+ {
+ "epoch": 7.794425087108014,
+ "grad_norm": 0.030088653787970543,
+ "learning_rate": 0.0005750410090362854,
+ "loss": 2.4312644004821777,
+ "step": 561
+ },
+ {
+ "epoch": 7.80836236933798,
+ "grad_norm": 0.03572467714548111,
+ "learning_rate": 0.0005749339281984761,
+ "loss": 2.429471969604492,
+ "step": 562
+ },
+ {
+ "epoch": 7.822299651567945,
+ "grad_norm": 0.0341353677213192,
+ "learning_rate": 0.0005748266282684124,
+ "loss": 2.4277775287628174,
+ "step": 563
+ },
+ {
+ "epoch": 7.83623693379791,
+ "grad_norm": 0.03292057290673256,
+ "learning_rate": 0.0005747191093325443,
+ "loss": 2.452070474624634,
+ "step": 564
+ },
+ {
+ "epoch": 7.850174216027875,
+ "grad_norm": 0.03879968449473381,
+ "learning_rate": 0.0005746113714774976,
+ "loss": 2.430136203765869,
+ "step": 565
+ },
+ {
+ "epoch": 7.86411149825784,
+ "grad_norm": 0.04012257233262062,
+ "learning_rate": 0.0005745034147900747,
+ "loss": 2.4385483264923096,
+ "step": 566
+ },
+ {
+ "epoch": 7.878048780487805,
+ "grad_norm": 0.0409339964389801,
+ "learning_rate": 0.0005743952393572544,
+ "loss": 2.452726125717163,
+ "step": 567
+ },
+ {
+ "epoch": 7.89198606271777,
+ "grad_norm": 0.044214967638254166,
+ "learning_rate": 0.0005742868452661918,
+ "loss": 2.4304707050323486,
+ "step": 568
+ },
+ {
+ "epoch": 7.905923344947735,
+ "grad_norm": 0.04643729701638222,
+ "learning_rate": 0.0005741782326042181,
+ "loss": 2.4338128566741943,
+ "step": 569
+ },
+ {
+ "epoch": 7.9198606271777,
+ "grad_norm": 0.05278981477022171,
+ "learning_rate": 0.0005740694014588403,
+ "loss": 2.439448595046997,
+ "step": 570
+ },
+ {
+ "epoch": 7.933797909407666,
+ "grad_norm": 0.05788639932870865,
+ "learning_rate": 0.0005739603519177419,
+ "loss": 2.4337964057922363,
+ "step": 571
+ },
+ {
+ "epoch": 7.947735191637631,
+ "grad_norm": 0.04502451792359352,
+ "learning_rate": 0.0005738510840687821,
+ "loss": 2.441267251968384,
+ "step": 572
+ },
+ {
+ "epoch": 7.961672473867596,
+ "grad_norm": 0.04129915311932564,
+ "learning_rate": 0.000573741597999996,
+ "loss": 2.43253231048584,
+ "step": 573
+ },
+ {
+ "epoch": 7.975609756097561,
+ "grad_norm": 0.04136184602975845,
+ "learning_rate": 0.0005736318937995947,
+ "loss": 2.4336652755737305,
+ "step": 574
+ },
+ {
+ "epoch": 7.989547038327526,
+ "grad_norm": 0.04168929532170296,
+ "learning_rate": 0.0005735219715559646,
+ "loss": 2.4467530250549316,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.02795771323144436,
+ "learning_rate": 0.0005734118313576683,
+ "loss": 1.8154387474060059,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 0.6265157461166382,
+ "eval_runtime": 40.9482,
+ "eval_samples_per_second": 59.636,
+ "eval_steps_per_second": 0.488,
+ "step": 576
+ },
+ {
+ "epoch": 8.013937282229966,
+ "grad_norm": 0.039702385663986206,
+ "learning_rate": 0.0005733014732934436,
+ "loss": 2.405975818634033,
+ "step": 577
+ },
+ {
+ "epoch": 8.02787456445993,
+ "grad_norm": 0.048223428428173065,
+ "learning_rate": 0.0005731908974522042,
+ "loss": 2.389124631881714,
+ "step": 578
+ },
+ {
+ "epoch": 8.041811846689896,
+ "grad_norm": 0.054057810455560684,
+ "learning_rate": 0.0005730801039230389,
+ "loss": 2.4132707118988037,
+ "step": 579
+ },
+ {
+ "epoch": 8.05574912891986,
+ "grad_norm": 0.06689552962779999,
+ "learning_rate": 0.0005729690927952123,
+ "loss": 2.412965774536133,
+ "step": 580
+ },
+ {
+ "epoch": 8.069686411149826,
+ "grad_norm": 0.07542029023170471,
+ "learning_rate": 0.0005728578641581637,
+ "loss": 2.4143495559692383,
+ "step": 581
+ },
+ {
+ "epoch": 8.08362369337979,
+ "grad_norm": 0.06956230849027634,
+ "learning_rate": 0.0005727464181015081,
+ "loss": 2.403184413909912,
+ "step": 582
+ },
+ {
+ "epoch": 8.097560975609756,
+ "grad_norm": 0.062457375228405,
+ "learning_rate": 0.0005726347547150357,
+ "loss": 2.415438175201416,
+ "step": 583
+ },
+ {
+ "epoch": 8.111498257839722,
+ "grad_norm": 0.06481153517961502,
+ "learning_rate": 0.0005725228740887117,
+ "loss": 2.408278465270996,
+ "step": 584
+ },
+ {
+ "epoch": 8.125435540069686,
+ "grad_norm": 0.06572942435741425,
+ "learning_rate": 0.0005724107763126761,
+ "loss": 2.402078151702881,
+ "step": 585
+ },
+ {
+ "epoch": 8.139372822299652,
+ "grad_norm": 0.0697183907032013,
+ "learning_rate": 0.0005722984614772442,
+ "loss": 2.4010097980499268,
+ "step": 586
+ },
+ {
+ "epoch": 8.153310104529616,
+ "grad_norm": 0.07151805609464645,
+ "learning_rate": 0.000572185929672906,
+ "loss": 2.4116389751434326,
+ "step": 587
+ },
+ {
+ "epoch": 8.167247386759582,
+ "grad_norm": 0.06399805098772049,
+ "learning_rate": 0.0005720731809903263,
+ "loss": 2.4083070755004883,
+ "step": 588
+ },
+ {
+ "epoch": 8.181184668989546,
+ "grad_norm": 0.057741884142160416,
+ "learning_rate": 0.000571960215520345,
+ "loss": 2.4084267616271973,
+ "step": 589
+ },
+ {
+ "epoch": 8.195121951219512,
+ "grad_norm": 0.05918360874056816,
+ "learning_rate": 0.0005718470333539757,
+ "loss": 2.4162850379943848,
+ "step": 590
+ },
+ {
+ "epoch": 8.209059233449477,
+ "grad_norm": 0.06101354956626892,
+ "learning_rate": 0.0005717336345824077,
+ "loss": 2.416327476501465,
+ "step": 591
+ },
+ {
+ "epoch": 8.222996515679442,
+ "grad_norm": 0.06537748873233795,
+ "learning_rate": 0.0005716200192970043,
+ "loss": 2.416635036468506,
+ "step": 592
+ },
+ {
+ "epoch": 8.236933797909408,
+ "grad_norm": 0.06058003753423691,
+ "learning_rate": 0.0005715061875893032,
+ "loss": 2.4121570587158203,
+ "step": 593
+ },
+ {
+ "epoch": 8.250871080139373,
+ "grad_norm": 0.06355198472738266,
+ "learning_rate": 0.0005713921395510166,
+ "loss": 2.4172019958496094,
+ "step": 594
+ },
+ {
+ "epoch": 8.264808362369338,
+ "grad_norm": 0.06816858053207397,
+ "learning_rate": 0.0005712778752740307,
+ "loss": 2.4226150512695312,
+ "step": 595
+ },
+ {
+ "epoch": 8.278745644599303,
+ "grad_norm": 0.05583881959319115,
+ "learning_rate": 0.0005711633948504066,
+ "loss": 2.4039454460144043,
+ "step": 596
+ },
+ {
+ "epoch": 8.292682926829269,
+ "grad_norm": 0.04999455437064171,
+ "learning_rate": 0.0005710486983723787,
+ "loss": 2.413280963897705,
+ "step": 597
+ },
+ {
+ "epoch": 8.306620209059233,
+ "grad_norm": 0.05040449649095535,
+ "learning_rate": 0.0005709337859323561,
+ "loss": 2.4149763584136963,
+ "step": 598
+ },
+ {
+ "epoch": 8.320557491289199,
+ "grad_norm": 0.04819140210747719,
+ "learning_rate": 0.0005708186576229218,
+ "loss": 2.4253149032592773,
+ "step": 599
+ },
+ {
+ "epoch": 8.334494773519165,
+ "grad_norm": 0.04608165845274925,
+ "learning_rate": 0.0005707033135368323,
+ "loss": 2.3984503746032715,
+ "step": 600
+ },
+ {
+ "epoch": 8.348432055749129,
+ "grad_norm": 0.04424264654517174,
+ "learning_rate": 0.0005705877537670184,
+ "loss": 2.414839267730713,
+ "step": 601
+ },
+ {
+ "epoch": 8.362369337979095,
+ "grad_norm": 0.04491661861538887,
+ "learning_rate": 0.0005704719784065846,
+ "loss": 2.4017510414123535,
+ "step": 602
+ },
+ {
+ "epoch": 8.376306620209059,
+ "grad_norm": 0.04609810188412666,
+ "learning_rate": 0.0005703559875488088,
+ "loss": 2.4047470092773438,
+ "step": 603
+ },
+ {
+ "epoch": 8.390243902439025,
+ "grad_norm": 0.04852808639407158,
+ "learning_rate": 0.0005702397812871429,
+ "loss": 2.4138975143432617,
+ "step": 604
+ },
+ {
+ "epoch": 8.404181184668989,
+ "grad_norm": 0.046814873814582825,
+ "learning_rate": 0.0005701233597152121,
+ "loss": 2.4012837409973145,
+ "step": 605
+ },
+ {
+ "epoch": 8.418118466898955,
+ "grad_norm": 0.04141746833920479,
+ "learning_rate": 0.0005700067229268154,
+ "loss": 2.415524959564209,
+ "step": 606
+ },
+ {
+ "epoch": 8.43205574912892,
+ "grad_norm": 0.04140830039978027,
+ "learning_rate": 0.0005698898710159245,
+ "loss": 2.4305288791656494,
+ "step": 607
+ },
+ {
+ "epoch": 8.445993031358885,
+ "grad_norm": 0.037624262273311615,
+ "learning_rate": 0.0005697728040766852,
+ "loss": 2.4011857509613037,
+ "step": 608
+ },
+ {
+ "epoch": 8.45993031358885,
+ "grad_norm": 0.0353207066655159,
+ "learning_rate": 0.0005696555222034162,
+ "loss": 2.4246044158935547,
+ "step": 609
+ },
+ {
+ "epoch": 8.473867595818815,
+ "grad_norm": 0.03710941970348358,
+ "learning_rate": 0.0005695380254906094,
+ "loss": 2.400528907775879,
+ "step": 610
+ },
+ {
+ "epoch": 8.487804878048781,
+ "grad_norm": 0.034571319818496704,
+ "learning_rate": 0.0005694203140329296,
+ "loss": 2.3888421058654785,
+ "step": 611
+ },
+ {
+ "epoch": 8.501742160278745,
+ "grad_norm": 0.0316578708589077,
+ "learning_rate": 0.000569302387925215,
+ "loss": 2.386704921722412,
+ "step": 612
+ },
+ {
+ "epoch": 8.515679442508711,
+ "grad_norm": 0.03137975186109543,
+ "learning_rate": 0.0005691842472624764,
+ "loss": 2.3930721282958984,
+ "step": 613
+ },
+ {
+ "epoch": 8.529616724738675,
+ "grad_norm": 0.03404382988810539,
+ "learning_rate": 0.0005690658921398977,
+ "loss": 2.4192662239074707,
+ "step": 614
+ },
+ {
+ "epoch": 8.543554006968641,
+ "grad_norm": 0.0336579792201519,
+ "learning_rate": 0.0005689473226528354,
+ "loss": 2.388643741607666,
+ "step": 615
+ },
+ {
+ "epoch": 8.557491289198607,
+ "grad_norm": 0.03660735860466957,
+ "learning_rate": 0.0005688285388968187,
+ "loss": 2.3898136615753174,
+ "step": 616
+ },
+ {
+ "epoch": 8.571428571428571,
+ "grad_norm": 0.036674171686172485,
+ "learning_rate": 0.0005687095409675499,
+ "loss": 2.396341323852539,
+ "step": 617
+ },
+ {
+ "epoch": 8.585365853658537,
+ "grad_norm": 0.041215281933546066,
+ "learning_rate": 0.000568590328960903,
+ "loss": 2.4068543910980225,
+ "step": 618
+ },
+ {
+ "epoch": 8.599303135888501,
+ "grad_norm": 0.036058180034160614,
+ "learning_rate": 0.0005684709029729253,
+ "loss": 2.408473014831543,
+ "step": 619
+ },
+ {
+ "epoch": 8.613240418118467,
+ "grad_norm": 0.032271627336740494,
+ "learning_rate": 0.0005683512630998361,
+ "loss": 2.391817092895508,
+ "step": 620
+ },
+ {
+ "epoch": 8.627177700348431,
+ "grad_norm": 0.03260865435004234,
+ "learning_rate": 0.000568231409438027,
+ "loss": 2.3981311321258545,
+ "step": 621
+ },
+ {
+ "epoch": 8.641114982578397,
+ "grad_norm": 0.031028274446725845,
+ "learning_rate": 0.0005681113420840619,
+ "loss": 2.3957371711730957,
+ "step": 622
+ },
+ {
+ "epoch": 8.655052264808361,
+ "grad_norm": 0.032890431582927704,
+ "learning_rate": 0.000567991061134677,
+ "loss": 2.3957810401916504,
+ "step": 623
+ },
+ {
+ "epoch": 8.668989547038327,
+ "grad_norm": 0.029823200777173042,
+ "learning_rate": 0.0005678705666867805,
+ "loss": 2.40175199508667,
+ "step": 624
+ },
+ {
+ "epoch": 8.682926829268293,
+ "grad_norm": 0.029909998178482056,
+ "learning_rate": 0.0005677498588374524,
+ "loss": 2.402420997619629,
+ "step": 625
+ },
+ {
+ "epoch": 8.696864111498257,
+ "grad_norm": 0.03211063891649246,
+ "learning_rate": 0.0005676289376839452,
+ "loss": 2.3948493003845215,
+ "step": 626
+ },
+ {
+ "epoch": 8.710801393728223,
+ "grad_norm": 0.03348059207201004,
+ "learning_rate": 0.0005675078033236827,
+ "loss": 2.393502712249756,
+ "step": 627
+ },
+ {
+ "epoch": 8.724738675958188,
+ "grad_norm": 0.03300804644823074,
+ "learning_rate": 0.0005673864558542605,
+ "loss": 2.3876824378967285,
+ "step": 628
+ },
+ {
+ "epoch": 8.738675958188153,
+ "grad_norm": 0.033442456275224686,
+ "learning_rate": 0.0005672648953734462,
+ "loss": 2.4107770919799805,
+ "step": 629
+ },
+ {
+ "epoch": 8.752613240418118,
+ "grad_norm": 0.03840150684118271,
+ "learning_rate": 0.000567143121979179,
+ "loss": 2.4036107063293457,
+ "step": 630
+ },
+ {
+ "epoch": 8.766550522648084,
+ "grad_norm": 0.041435301303863525,
+ "learning_rate": 0.0005670211357695693,
+ "loss": 2.387625217437744,
+ "step": 631
+ },
+ {
+ "epoch": 8.78048780487805,
+ "grad_norm": 0.042547907680273056,
+ "learning_rate": 0.0005668989368428994,
+ "loss": 2.3926758766174316,
+ "step": 632
+ },
+ {
+ "epoch": 8.794425087108014,
+ "grad_norm": 0.033522650599479675,
+ "learning_rate": 0.0005667765252976227,
+ "loss": 2.3900928497314453,
+ "step": 633
+ },
+ {
+ "epoch": 8.80836236933798,
+ "grad_norm": 0.03421244025230408,
+ "learning_rate": 0.0005666539012323639,
+ "loss": 2.403102397918701,
+ "step": 634
+ },
+ {
+ "epoch": 8.822299651567944,
+ "grad_norm": 0.03918502479791641,
+ "learning_rate": 0.0005665310647459189,
+ "loss": 2.403184175491333,
+ "step": 635
+ },
+ {
+ "epoch": 8.83623693379791,
+ "grad_norm": 0.03905736654996872,
+ "learning_rate": 0.0005664080159372551,
+ "loss": 2.3944578170776367,
+ "step": 636
+ },
+ {
+ "epoch": 8.850174216027874,
+ "grad_norm": 0.042079947888851166,
+ "learning_rate": 0.0005662847549055107,
+ "loss": 2.40266752243042,
+ "step": 637
+ },
+ {
+ "epoch": 8.86411149825784,
+ "grad_norm": 0.04590294137597084,
+ "learning_rate": 0.0005661612817499947,
+ "loss": 2.4020891189575195,
+ "step": 638
+ },
+ {
+ "epoch": 8.878048780487806,
+ "grad_norm": 0.04212665557861328,
+ "learning_rate": 0.0005660375965701872,
+ "loss": 2.392502784729004,
+ "step": 639
+ },
+ {
+ "epoch": 8.89198606271777,
+ "grad_norm": 0.03891785070300102,
+ "learning_rate": 0.0005659136994657392,
+ "loss": 2.394350528717041,
+ "step": 640
+ },
+ {
+ "epoch": 8.905923344947736,
+ "grad_norm": 0.03955228254199028,
+ "learning_rate": 0.0005657895905364723,
+ "loss": 2.3940305709838867,
+ "step": 641
+ },
+ {
+ "epoch": 8.9198606271777,
+ "grad_norm": 0.042912617325782776,
+ "learning_rate": 0.0005656652698823788,
+ "loss": 2.392199993133545,
+ "step": 642
+ },
+ {
+ "epoch": 8.933797909407666,
+ "grad_norm": 0.04031229391694069,
+ "learning_rate": 0.0005655407376036217,
+ "loss": 2.3964648246765137,
+ "step": 643
+ },
+ {
+ "epoch": 8.94773519163763,
+ "grad_norm": 0.03461204469203949,
+ "learning_rate": 0.0005654159938005345,
+ "loss": 2.401916027069092,
+ "step": 644
+ },
+ {
+ "epoch": 8.961672473867596,
+ "grad_norm": 0.03794260695576668,
+ "learning_rate": 0.0005652910385736207,
+ "loss": 2.4065444469451904,
+ "step": 645
+ },
+ {
+ "epoch": 8.975609756097562,
+ "grad_norm": 0.038141801953315735,
+ "learning_rate": 0.0005651658720235548,
+ "loss": 2.38955020904541,
+ "step": 646
+ },
+ {
+ "epoch": 8.989547038327526,
+ "grad_norm": 0.036538828164339066,
+ "learning_rate": 0.0005650404942511812,
+ "loss": 2.3973941802978516,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.028448769822716713,
+ "learning_rate": 0.0005649149053575141,
+ "loss": 1.7881890535354614,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 0.6202531456947327,
+ "eval_runtime": 40.4352,
+ "eval_samples_per_second": 60.393,
+ "eval_steps_per_second": 0.495,
+ "step": 648
+ },
+ {
+ "epoch": 9.013937282229966,
+ "grad_norm": 0.043303798884153366,
+ "learning_rate": 0.0005647891054437386,
+ "loss": 2.355170249938965,
+ "step": 649
+ },
+ {
+ "epoch": 9.02787456445993,
+ "grad_norm": 0.05378527566790581,
+ "learning_rate": 0.0005646630946112096,
+ "loss": 2.3758740425109863,
+ "step": 650
+ },
+ {
+ "epoch": 9.041811846689896,
+ "grad_norm": 0.05362150818109512,
+ "learning_rate": 0.0005645368729614512,
+ "loss": 2.361086845397949,
+ "step": 651
+ },
+ {
+ "epoch": 9.05574912891986,
+ "grad_norm": 0.054346974939107895,
+ "learning_rate": 0.0005644104405961582,
+ "loss": 2.3574094772338867,
+ "step": 652
+ },
+ {
+ "epoch": 9.069686411149826,
+ "grad_norm": 0.053007930517196655,
+ "learning_rate": 0.0005642837976171949,
+ "loss": 2.3493103981018066,
+ "step": 653
+ },
+ {
+ "epoch": 9.08362369337979,
+ "grad_norm": 0.06041328236460686,
+ "learning_rate": 0.0005641569441265952,
+ "loss": 2.3617541790008545,
+ "step": 654
+ },
+ {
+ "epoch": 9.097560975609756,
+ "grad_norm": 0.06779699772596359,
+ "learning_rate": 0.0005640298802265626,
+ "loss": 2.372866630554199,
+ "step": 655
+ },
+ {
+ "epoch": 9.111498257839722,
+ "grad_norm": 0.0699085146188736,
+ "learning_rate": 0.0005639026060194704,
+ "loss": 2.350039005279541,
+ "step": 656
+ },
+ {
+ "epoch": 9.125435540069686,
+ "grad_norm": 0.077919602394104,
+ "learning_rate": 0.0005637751216078609,
+ "loss": 2.3763508796691895,
+ "step": 657
+ },
+ {
+ "epoch": 9.139372822299652,
+ "grad_norm": 0.07544098049402237,
+ "learning_rate": 0.0005636474270944461,
+ "loss": 2.3677217960357666,
+ "step": 658
+ },
+ {
+ "epoch": 9.153310104529616,
+ "grad_norm": 0.08149590343236923,
+ "learning_rate": 0.0005635195225821072,
+ "loss": 2.36328387260437,
+ "step": 659
+ },
+ {
+ "epoch": 9.167247386759582,
+ "grad_norm": 0.08277004212141037,
+ "learning_rate": 0.0005633914081738945,
+ "loss": 2.3789758682250977,
+ "step": 660
+ },
+ {
+ "epoch": 9.181184668989546,
+ "grad_norm": 0.07791072130203247,
+ "learning_rate": 0.0005632630839730275,
+ "loss": 2.376584529876709,
+ "step": 661
+ },
+ {
+ "epoch": 9.195121951219512,
+ "grad_norm": 0.07431668043136597,
+ "learning_rate": 0.0005631345500828946,
+ "loss": 2.3792600631713867,
+ "step": 662
+ },
+ {
+ "epoch": 9.209059233449477,
+ "grad_norm": 0.06893738359212875,
+ "learning_rate": 0.0005630058066070533,
+ "loss": 2.354691505432129,
+ "step": 663
+ },
+ {
+ "epoch": 9.222996515679442,
+ "grad_norm": 0.07751812785863876,
+ "learning_rate": 0.0005628768536492299,
+ "loss": 2.388519763946533,
+ "step": 664
+ },
+ {
+ "epoch": 9.236933797909408,
+ "grad_norm": 0.07796601951122284,
+ "learning_rate": 0.0005627476913133193,
+ "loss": 2.3600292205810547,
+ "step": 665
+ },
+ {
+ "epoch": 9.250871080139373,
+ "grad_norm": 0.07397744804620743,
+ "learning_rate": 0.0005626183197033856,
+ "loss": 2.383859157562256,
+ "step": 666
+ },
+ {
+ "epoch": 9.264808362369338,
+ "grad_norm": 0.06795241683721542,
+ "learning_rate": 0.0005624887389236609,
+ "loss": 2.3526296615600586,
+ "step": 667
+ },
+ {
+ "epoch": 9.278745644599303,
+ "grad_norm": 0.06839273124933243,
+ "learning_rate": 0.0005623589490785462,
+ "loss": 2.3622193336486816,
+ "step": 668
+ },
+ {
+ "epoch": 9.292682926829269,
+ "grad_norm": 0.06846648454666138,
+ "learning_rate": 0.0005622289502726108,
+ "loss": 2.3708081245422363,
+ "step": 669
+ },
+ {
+ "epoch": 9.306620209059233,
+ "grad_norm": 0.06907503306865692,
+ "learning_rate": 0.0005620987426105925,
+ "loss": 2.364431381225586,
+ "step": 670
+ },
+ {
+ "epoch": 9.320557491289199,
+ "grad_norm": 0.05993546545505524,
+ "learning_rate": 0.000561968326197397,
+ "loss": 2.375314474105835,
+ "step": 671
+ },
+ {
+ "epoch": 9.334494773519165,
+ "grad_norm": 0.05129535496234894,
+ "learning_rate": 0.0005618377011380988,
+ "loss": 2.3689067363739014,
+ "step": 672
+ },
+ {
+ "epoch": 9.348432055749129,
+ "grad_norm": 0.05310920625925064,
+ "learning_rate": 0.0005617068675379398,
+ "loss": 2.3670191764831543,
+ "step": 673
+ },
+ {
+ "epoch": 9.362369337979095,
+ "grad_norm": 0.05011405423283577,
+ "learning_rate": 0.0005615758255023306,
+ "loss": 2.372076988220215,
+ "step": 674
+ },
+ {
+ "epoch": 9.376306620209059,
+ "grad_norm": 0.04965347424149513,
+ "learning_rate": 0.0005614445751368491,
+ "loss": 2.3823747634887695,
+ "step": 675
+ },
+ {
+ "epoch": 9.390243902439025,
+ "grad_norm": 0.045770980417728424,
+ "learning_rate": 0.0005613131165472415,
+ "loss": 2.360696792602539,
+ "step": 676
+ },
+ {
+ "epoch": 9.404181184668989,
+ "grad_norm": 0.05061430484056473,
+ "learning_rate": 0.0005611814498394216,
+ "loss": 2.3769307136535645,
+ "step": 677
+ },
+ {
+ "epoch": 9.418118466898955,
+ "grad_norm": 0.05156691372394562,
+ "learning_rate": 0.0005610495751194708,
+ "loss": 2.3763058185577393,
+ "step": 678
+ },
+ {
+ "epoch": 9.43205574912892,
+ "grad_norm": 0.04398881644010544,
+ "learning_rate": 0.0005609174924936384,
+ "loss": 2.3687405586242676,
+ "step": 679
+ },
+ {
+ "epoch": 9.445993031358885,
+ "grad_norm": 0.04115891084074974,
+ "learning_rate": 0.0005607852020683409,
+ "loss": 2.3599789142608643,
+ "step": 680
+ },
+ {
+ "epoch": 9.45993031358885,
+ "grad_norm": 0.038642384111881256,
+ "learning_rate": 0.0005606527039501621,
+ "loss": 2.3829524517059326,
+ "step": 681
+ },
+ {
+ "epoch": 9.473867595818815,
+ "grad_norm": 0.03955807536840439,
+ "learning_rate": 0.0005605199982458535,
+ "loss": 2.3717896938323975,
+ "step": 682
+ },
+ {
+ "epoch": 9.487804878048781,
+ "grad_norm": 0.0387716107070446,
+ "learning_rate": 0.0005603870850623338,
+ "loss": 2.352104425430298,
+ "step": 683
+ },
+ {
+ "epoch": 9.501742160278745,
+ "grad_norm": 0.035999514162540436,
+ "learning_rate": 0.0005602539645066884,
+ "loss": 2.3635125160217285,
+ "step": 684
+ },
+ {
+ "epoch": 9.515679442508711,
+ "grad_norm": 0.03553462028503418,
+ "learning_rate": 0.0005601206366861706,
+ "loss": 2.3691720962524414,
+ "step": 685
+ },
+ {
+ "epoch": 9.529616724738675,
+ "grad_norm": 0.03732144460082054,
+ "learning_rate": 0.0005599871017081999,
+ "loss": 2.3733935356140137,
+ "step": 686
+ },
+ {
+ "epoch": 9.543554006968641,
+ "grad_norm": 0.034763403236866,
+ "learning_rate": 0.0005598533596803631,
+ "loss": 2.3709394931793213,
+ "step": 687
+ },
+ {
+ "epoch": 9.557491289198607,
+ "grad_norm": 0.034646645188331604,
+ "learning_rate": 0.0005597194107104137,
+ "loss": 2.3806018829345703,
+ "step": 688
+ },
+ {
+ "epoch": 9.571428571428571,
+ "grad_norm": 0.033737875521183014,
+ "learning_rate": 0.000559585254906272,
+ "loss": 2.371210813522339,
+ "step": 689
+ },
+ {
+ "epoch": 9.585365853658537,
+ "grad_norm": 0.03646186366677284,
+ "learning_rate": 0.000559450892376025,
+ "loss": 2.3905677795410156,
+ "step": 690
+ },
+ {
+ "epoch": 9.599303135888501,
+ "grad_norm": 0.039898719638586044,
+ "learning_rate": 0.000559316323227926,
+ "loss": 2.3705878257751465,
+ "step": 691
+ },
+ {
+ "epoch": 9.613240418118467,
+ "grad_norm": 0.041938189417123795,
+ "learning_rate": 0.000559181547570395,
+ "loss": 2.356766939163208,
+ "step": 692
+ },
+ {
+ "epoch": 9.627177700348431,
+ "grad_norm": 0.0372963547706604,
+ "learning_rate": 0.0005590465655120183,
+ "loss": 2.3734800815582275,
+ "step": 693
+ },
+ {
+ "epoch": 9.641114982578397,
+ "grad_norm": 0.035154446959495544,
+ "learning_rate": 0.0005589113771615486,
+ "loss": 2.3585588932037354,
+ "step": 694
+ },
+ {
+ "epoch": 9.655052264808361,
+ "grad_norm": 0.03412827104330063,
+ "learning_rate": 0.0005587759826279046,
+ "loss": 2.3630642890930176,
+ "step": 695
+ },
+ {
+ "epoch": 9.668989547038327,
+ "grad_norm": 0.03843880817294121,
+ "learning_rate": 0.0005586403820201713,
+ "loss": 2.3606934547424316,
+ "step": 696
+ },
+ {
+ "epoch": 9.682926829268293,
+ "grad_norm": 0.031355664134025574,
+ "learning_rate": 0.0005585045754475995,
+ "loss": 2.3663330078125,
+ "step": 697
+ },
+ {
+ "epoch": 9.696864111498257,
+ "grad_norm": 0.0323006808757782,
+ "learning_rate": 0.0005583685630196064,
+ "loss": 2.366168975830078,
+ "step": 698
+ },
+ {
+ "epoch": 9.710801393728223,
+ "grad_norm": 0.0335552878677845,
+ "learning_rate": 0.0005582323448457744,
+ "loss": 2.357715129852295,
+ "step": 699
+ },
+ {
+ "epoch": 9.724738675958188,
+ "grad_norm": 0.031565241515636444,
+ "learning_rate": 0.0005580959210358523,
+ "loss": 2.360602378845215,
+ "step": 700
+ },
+ {
+ "epoch": 9.738675958188153,
+ "grad_norm": 0.034486074000597,
+ "learning_rate": 0.0005579592916997542,
+ "loss": 2.369016408920288,
+ "step": 701
+ },
+ {
+ "epoch": 9.752613240418118,
+ "grad_norm": 0.03649681806564331,
+ "learning_rate": 0.0005578224569475599,
+ "loss": 2.3676323890686035,
+ "step": 702
+ },
+ {
+ "epoch": 9.766550522648084,
+ "grad_norm": 0.0355026014149189,
+ "learning_rate": 0.0005576854168895147,
+ "loss": 2.362356662750244,
+ "step": 703
+ },
+ {
+ "epoch": 9.78048780487805,
+ "grad_norm": 0.03701469302177429,
+ "learning_rate": 0.0005575481716360292,
+ "loss": 2.35752534866333,
+ "step": 704
+ },
+ {
+ "epoch": 9.794425087108014,
+ "grad_norm": 0.03648044168949127,
+ "learning_rate": 0.0005574107212976796,
+ "loss": 2.3521745204925537,
+ "step": 705
+ },
+ {
+ "epoch": 9.80836236933798,
+ "grad_norm": 0.03555280342698097,
+ "learning_rate": 0.000557273065985207,
+ "loss": 2.347022771835327,
+ "step": 706
+ },
+ {
+ "epoch": 9.822299651567944,
+ "grad_norm": 0.03129667788743973,
+ "learning_rate": 0.0005571352058095179,
+ "loss": 2.3691742420196533,
+ "step": 707
+ },
+ {
+ "epoch": 9.83623693379791,
+ "grad_norm": 0.0333794429898262,
+ "learning_rate": 0.0005569971408816838,
+ "loss": 2.343442916870117,
+ "step": 708
+ },
+ {
+ "epoch": 9.850174216027874,
+ "grad_norm": 0.03256125748157501,
+ "learning_rate": 0.0005568588713129409,
+ "loss": 2.3567285537719727,
+ "step": 709
+ },
+ {
+ "epoch": 9.86411149825784,
+ "grad_norm": 0.03400867432355881,
+ "learning_rate": 0.0005567203972146906,
+ "loss": 2.3687191009521484,
+ "step": 710
+ },
+ {
+ "epoch": 9.878048780487806,
+ "grad_norm": 0.03227195143699646,
+ "learning_rate": 0.000556581718698499,
+ "loss": 2.364328145980835,
+ "step": 711
+ },
+ {
+ "epoch": 9.89198606271777,
+ "grad_norm": 0.031485479325056076,
+ "learning_rate": 0.000556442835876097,
+ "loss": 2.3622374534606934,
+ "step": 712
+ },
+ {
+ "epoch": 9.905923344947736,
+ "grad_norm": 0.031078679487109184,
+ "learning_rate": 0.0005563037488593799,
+ "loss": 2.3633031845092773,
+ "step": 713
+ },
+ {
+ "epoch": 9.9198606271777,
+ "grad_norm": 0.0335998609662056,
+ "learning_rate": 0.0005561644577604074,
+ "loss": 2.3585972785949707,
+ "step": 714
+ },
+ {
+ "epoch": 9.933797909407666,
+ "grad_norm": 0.033543020486831665,
+ "learning_rate": 0.000556024962691404,
+ "loss": 2.3478875160217285,
+ "step": 715
+ },
+ {
+ "epoch": 9.94773519163763,
+ "grad_norm": 0.03368858993053436,
+ "learning_rate": 0.0005558852637647584,
+ "loss": 2.370272397994995,
+ "step": 716
+ },
+ {
+ "epoch": 9.961672473867596,
+ "grad_norm": 0.03240448608994484,
+ "learning_rate": 0.0005557453610930235,
+ "loss": 2.371398687362671,
+ "step": 717
+ },
+ {
+ "epoch": 9.975609756097562,
+ "grad_norm": 0.0316084623336792,
+ "learning_rate": 0.0005556052547889161,
+ "loss": 2.36739444732666,
+ "step": 718
+ },
+ {
+ "epoch": 9.989547038327526,
+ "grad_norm": 0.0325324572622776,
+ "learning_rate": 0.0005554649449653177,
+ "loss": 2.3624932765960693,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.025275887921452522,
+ "learning_rate": 0.0005553244317352731,
+ "loss": 1.7544682025909424,
+ "step": 720
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.0776560164274176e+17,
+ "train_batch_size": 64,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/mt5-ssl/checkpoint-720/training_args.bin b/runs/mt5-ssl/checkpoint-720/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..5b4b00290c5cc346fb17bc1f2462108857f0fa38
--- /dev/null
+++ b/runs/mt5-ssl/checkpoint-720/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5709a4c4a39fd15af555606dc0c6fc500314d925d656b6d8229a416e911ad6b4
+size 5201